PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
4
5
4
5
3.8
置信度
创新性2.8
质量2.8
清晰度2.5
重要性2.5
NeurIPS 2025

AcuRank: Uncertainty-Aware Adaptive Computation for Listwise Reranking

OpenReviewPDF
提交: 2025-05-12更新: 2025-10-29
TL;DR

AcuRank adaptively reranks documents using uncertainty, improving ranking accuracy with fewer LLM calls.

摘要

关键词
Listwise RerankingLarge Language ModelsAdaptive ComputationUncertainty EstimationInference Efficiency

评审与讨论

审稿意见
4

The existing methods for listwise reranking using large language models are computationally expensive. While there have been various attempts to address these issues, they still suffer from problems of rigidity and lack of adaptability. To overcome these challenges, this paper proposes AcuRank, an uncertainty-aware reranking method, which utilizes the TrueSkill Bayesian rating system for reranking.

优缺点分析

Strengths

  • The paper is well written and easy to follow.
  • Using uncertainty as for the reranking metric is an interesting attempt.
  • Experimental results prove the efficacy of the proposed method.

Weaknesses

  • It is not trivial to understand the motivation for setting the initialization variance to σ2/3\sigma^2/3. It would be helpful to have either an ablation experiment or more motivation to clarify this.
  • It would be great to have a qualitative example showing how the predictive mean and ranking of the retrieved documents change during the Bayesian update. For instance, documents that initially had high scores but were uncertain might have their scores decrease and uncertainty reduce, leading to a lower rank after the update, or vice versa.
  • It would be helpful for readers if there exist the ablation experiments on the uncertainty threshold.
  • It would be helpful to analyze how the scale of the updated mean and the number of updates required for the final update change as the number of retrieved documents, nn, varies. This would provide a clearer understanding of how the method operates as nn changes.
  • Let’s consider a document set D1,,Dn{\mathcal{D}_1, \ldots, \mathcal{D}_n}. After reranking this set, we obtain the result D^1,,D^n{\hat{\mathcal{D}}_1, \ldots, \hat{\mathcal{D}}_n}. Now, suppose a new set of documents Dˉ1,,Dˉm{\bar{\mathcal{D}}_1, \ldots, \bar{\mathcal{D}}_m} is added to the collection. The task is to examine whether the reranked result of the combined set D1,,Dn,Dˉ1,,Dˉm{\mathcal{D}_1, \ldots, \mathcal{D}_n, \bar{\mathcal{D}}_1, \ldots, \bar{\mathcal{D}}_m^*} preserves the relative order of the original documents ${\mathcal{D}_1, \ldots, \mathcal{D}_n}$ as ${\hat{\mathcal{D}}_1, \ldots, \hat{\mathcal{D}}_n}$. An experiment focusing on this could help verify whether the method maintains the relative ranking robustly, regardless of the composition of the document set. This would ensure that the method is consistent in preserving the order of the original documents even when new documents are added.

问题

See the Weaknesses section.

局限性

See the weaknesses section.

I am willing to increase the rating based on the authors response on my questions.

最终评判理由

They provide additional explanations and experiments which are reasonable to increase the score.

格式问题

Nothing

作者回复

Thank you for providing valuable comments to improve our paper! We are especially encouraged by the recognition of our adaptive computation strategy in listwise reranking. We hope our comments are sufficient to answer the questions raised and that you consider raising the overall rating of our work. We would greatly appreciate it if you could review our comments and inform us if any further discussions or clarification is required.

1. Variance Initialization

  • We appreciate the reviewer’s question. As described in Supplementary Section 2.4, we provide motivation for our choice of initializing each document’s relevance estimate with a standard deviation proportional to its retrieval score (σi=μi/3\sigma_i = \mu_i / 3). This setting reflects the idea that documents with higher initial scores may carry greater uncertainty due to potential overestimation. The 1/3 ratio is adopted from the original TrueSkill configuration and maintains a statistically meaningful spread around the prior mean.
  • To further support this design, we conduct an ablation study comparing three initialization variants: (1) a fixed uncertainty (σi=8\sigma_i = 8), (2) a more conservative variant (μi/2\mu_i / 2), and (3) our default setting (μi/3\mu_i / 3). As shown in the table below, the default configuration performs best overall, though all variants preserve the same qualitative trends. This confirms that AcuRank is robust to the specific choice of initialization and supports the effectiveness of our design.

Table 8: Ablations on initializing variance (BM25 top-100, RankZephyr)

σ\sigmaDL19DL20DL21DL22DL23DL-HCOVIDNFCSignalNewsR04ToucheDBPScifAvg.# Calls
8 (fixed)74.172.670.551.346.939.585.536.931.152.556.233.345.176.455.113.1
μ/2\mu / 274.171.370.751.846.239.485.237.432.052.656.337.346.075.755.416.4
μ/3\mu / 3 (AcuRank)74.271.870.352.047.039.485.337.231.853.956.636.546.075.355.519.2

2. Qualitative Example of Bayesian Update

  • We appreciate the reviewer’s request for a qualitative walk-through.
  • Setup. The table below shows the trace for the first query of TREC-DL19 as AcuRank iterates five times over 50 candidate passages (n = 50). We highlight two contrasting passages to make the dynamics clear.
  • Over-estimated early leader (pid 6337909): On the first iteration this non-relevant passage holds a slightly higher mean score than the gold passage (μ\mu = 13.77 > 13.44). After one unfavourable comparison its mean collapses, confidence tightens, and because subsequent comparisons are unnecessary, the passage is frozen at rank 33.
  • Initally uncertain passage that surges (pid 4647186, gold). Starting lower than 6337909, this passage wins multiple pairwise contests. Each victory pushes its mean up and its σ down until it confidently enters the top-10 (rank 9).
  • The key take-away is that, uncertain winners can surge, over-estimated items are pruned, and edits stop once the confidence is high, saving compute.

Table 9. Per-iteration trace for Gold-relevant pid 4647186

PassμσRank
113.442.9620
215.482.4614
317.062.1510
418.271.9410
519.021.789 ← final

Table 10. Per-iteration trace for Non-relevant pid 6337909

PassμσRank
113.773.1419
29.502.6633
39.502.6633
49.502.6633
59.502.6633 ← final

Net movement:

  • Gold passage climbs 11 rank out of 50 candidates (20 → 9).
  • Non-gold passage drops 14 ranks out of 50 candidates(19 → 33).

This examplifies how AcuRank's Bayesian evidence accumulation boosts performance while saving LLM calls by halting once posterior uncertainty is low.

3. Ablations on Uncertainty Threshold

  • Please see our detailed response to Reviewer SkZp under the "Choice of Uncertainty Thresholds" section. We had already explored the impact of different uncertainty thresholds through our original configurations (AcuRank-H, AcuRank-HH, and AcuRank-9), which represent varying trade-offs between accuracy and efficiency. In response to the reviewer’s suggestion, we additionally conducted ablation experiments by varying the thresholds ϵ\epsilon and τ\tau. These results exhibit smooth and monotonic performance-efficiency trade-offs and confirm the robustness of AcuRank across a wide range of settings.

4. Impact on Update Dynamics over Varying n

  • We thank the reviewer for this insightful suggestion. While our current experiments fix nn based on the first-stage retriever (e.g., top-100 or top-1000), we discuss here how varying nn could affect the TrueSkill update process in AcuRank.
  • In general, increasing nn leads to a larger candidate pool and potentially more reranker calls. However, AcuRank's adaptive computation adjusts the number of updates based on the model's uncertainty. Even as nn increases, AcuRank still focuses reranker calls on a relatively small subset of uncertain documents, which can result in sublinear growth in total computation.
  • We conduct and report additional experiment upon reviewer's request. With varying nn to 50, 100, 200, and 1000, we measure (1) the average number of updates unitl convergence and (2) the average updated mean (μ\mu) among the top-100 candidates. (top-50 for n=50n = 50).
  • From the result tables shown below, we find that the adaptive behavior of AcuRank helps maintain scalability (lower # calls on increased nn) compared with the sliding window baselines, especially as the the candidate pool size nn grows. For example, when nn increases from 100 to 1000 (10x), the average number of calls of AcuRank only triples (68.6 / 18.7 = 3.7x). In contrast, the sliding windows baselines increase almost linearly (e.g., SW-1: 99 / 9 = 11x). Also, we notice a mild increase in the average updated μ\mu scores as nn increases, which is expected since we measured over the top-100.
  • We appreciate the reviewer for highlighting the insightful direction of analysis, and we will include those empirically analyzed results in the revised version.

Table 11: AcuRank, # Calls to Converge (rightmost is for SW-1)

nnDL19DL20COVIDNewsAvgSW-1 Avg
5012.613.214.312.013.04.0
10018.316.321.818.518.79.0
20026.323.630.625.528.019.0
100067.865.672.868.068.699.0

Table 12: AcuRank, μ\mu Mean (max top-100)

nnDL19DL20COVIDNewsAvgSW-1 Avg
5010.811.610.012.511.2-
1009.910.59.411.510.3-
20011.512.211.013.412.2-
100012.012.811.614.112.6-

5. Stability of Rankings after Document Addition

  • We thank the reviewer for this interesting suggestion. AcuRank is currently designed to identify the top-kk most relevant documents through adaptive reranking. While it can, in principle, produce a full ordering over the candidate set based on estimated relevance scores, its computational focus is on accurately ranking the most promising candidates. As a result, documents identified early as unlikely to belong in the top-kk typically receive fewer updates, and their relative ordering may be less reliable.
  • To examine the stability of ranking under document addition, we conduct experiments using a document set D\mathcal{D} and its extended version D+\mathcal{D}^{+}, where D+D\mathcal{D}^{+} \supseteq \mathcal{D}. Specifically, we use the top-50 and top-100 documents from BM25 first-stage retrieval as D\mathcal{D} and D+\mathcal{D}^{+}, respectively (D=50|\mathcal{D}| = 50, D+=100|\mathcal{D}^{+}| = 100). We apply listwise reranking (AcuRank and SW-1 as a baseline) to both sets independently. Then, for a subset SD\mathcal{S} \subset \mathcal{D}, we compute the proportion of preserved pairwise relative rankings across the two reranked results.
  • We consider two variants of S\mathcal{S}: (1) the full set D\mathcal{D}, which corresponds to the global ranking stability considered by the reviewer, and (2) the subset of relevant documents within D\mathcal{D} based on gold labels. The latter better reflects our practical focus: since AcuRank prioritizes ranking accuracy for top-kk candidates, documents deemed irrelevant early on are unlikely to receive sufficient updates, and their mutual ordering may not be reliable. In both cases, AcuRank consistently preserves relative rankings more robustly than SW-1, as shown in the table below.

Table 13: Preserved ratio of pairwise relative ranking

S\mathcal{S}MethodDL19DL20TREC-CovidNews
D\mathcal{D}AcuRank86.6%88.7%85.6%86.8%
D\mathcal{D}SW-185.8%84.8%83.7%83.6%
GoldAcuRank87.7%88.2%85.3%87.9%
GoldSW-188.7%86.2%81.9%86.0%
评论

Thank you to the authors for the detailed explanations and additional experiments. Overall, I am satisfied with the responses, and I will consider adjusting my score if one final question is addressed.

My question regarding Q1 was originally about why the value '3' is used as the divisor. As shown in the authors’ experiments, dividing by '3' appears to work better than dividing by '2', but I am curious as to why this particular value of '3' should lead to better performance. I understand that the authors may have followed the convention established in TrueSkill, and it is possible they may not have a definitive reason, which I think is completely understandable. I’m simply interested in hearing the authors’ thoughts on this matter, and if they can offer any explanation or intuition behind it, I would appreciate it.

评论

We thank the reviewer for the opportunity to clarify this point. Below we provide our main intuitions behind the three-sigma rule; initializing the prior standard deviation as σ=μ/3\sigma = \mu/3 (rather than, say, μ/2\mu/2).

1. A Conservative Yet Flexible Starting Point

  • Ensures non-negative displayed scores for semantic compatibility. In TrueSkill, the displayed score is computed as R=μ3σR = \mu - 3\sigma. Setting σ=μ/3\sigma = \mu/3 results in R=0R = 0, ensuring a conservative lower bound without making overly pessimistic assumptions. Since relevance scores are often expected to be non-negative, this initialization preserves semantic consistency and avoids issues with downstream use of these scores (e.g., in log-based metrics or loss functions, where negative values may be undefined or unstable).

  • Allows generous headroom for updates. Statistically, [μiμi,μi+μi]=[0,2μi][\mu_i-\mu_i, \mu_i+\mu_i] = [0, 2\mu_i] corresponds to [μi±3σi][\mu_i \pm 3\sigma_i], since 3σi=μi3\sigma_i = \mu_i (Supplementary L.66). This gives the model freedom to significantly revise its belief upward if warranted by reranker signals, allowing the final estimated score to be up to twice the initial retrieval value, striking a balance between caution and flexibility.

2. A Statistically Grounded Default

  • Three-sigma covers almost all plausible values in a normal distribution. In Gaussian distributions, ±3σ\sigma encloses 99.7 % of probability mass (Supplementary L.67). This is commonly interpreted as "a (very) conservative estimate for your skill. You’re probably better than this conservative estimate, but you’re most likely not worse than this value" [1]. In other words, the "system is 99% sure that the player's skill is actually higher than what is displayed as their rank." [3] Accordingly, the three-sigma rule is widely used in quality control, anomaly detection, and the original TrueSkill paper [2]. It incorporates the retrieval signal while leaving room for the re‑ranking stage to meaningfully revise the belief (Supplementary L.69).
  • Avoids over- or under-confidence. Dividing by 2 instead of 3 would increase the prior variance, making the initialization overly uncertain and diminishing the value of the retrieval signal. Conversely, dividing by 4 would make the prior overly confident, reducing the model’s ability to adjust based on reranker feedback. The 3:1 ratio provides a principled middle ground.

3. Empirical Support and Established Practice

  • Adopted in large-scale systems and libraries. The 3:1 μ:σ ratio is commonly used in Xbox Live, multiple e-sports ranking systems, and the open-source trueskill Python library. [3] These systems have been tested on millions of matches and scenarios, indicating that this initialization is a robust and effective default across very diverse score distributions, offering a strong balance between fast convergence and robustness to noise.

In short, we adopt the σ=μ/3\sigma = \mu/3 initialization because it reflects a well-established statistical heuristic that provides a conservative yet flexible prior. It maintains semantic consistency, enables scalable updates, and has a strong track record of practical success. We hope this explanation sufficiently addresses the reviewer’s question. We will incorporate a concise explanation of this rationale in the final version.

[1] Moser, J. “Computing Your Skill.” Technical blog post, Moserware, 2010

[2] Herbrich, R., Minka, T., & Graepel, T. (2007). TrueSkill™: A Bayesian skill rating system.

[3] Wikipedia, TrueSkill (accessed 2025).

评论

Thank you for the additional comments. Please update the paper base on all the discussions. I'll update the score from 3 to 4.

审稿意见
5

This paper proposes an approach for information retrieval via second-stage reranking using LLMs. The limited context length of LLMs necessitates breaking down large first-stage candidate sets into small batches, and the high inference cost motivates minimizing the number of LLM calls. Existing approaches like sliding-window and tournament-style methods apply a fixed computational budget. The proposed approach models uncertainty in document relevance scores using the TrueSkill model (represents relevance not as a single point value, but as a Gaussian distribution) and uses it to decide which documents need to be reranked by the LLM. This allows dynamic choice of documents for comparison and early stopping when the uncertainty for top ranked documents is small. The approach shows promise in terms of accuracy-efficiency trade-off on several retrieval benchmarks (TREC-DL and BEIR).

优缺点分析

Strengths:

  • The proposed approach is principled and well justified.
  • Using relevance uncertainty to select documents for reranking with TrueSkill is an elegant solution.
  • The approach serves as a meta-framework using the LLM as a black-box reranker, making it a practical and broadly applicable approach.
  • Empirical evaluation shows compelling results in terms of accuracy-efficiency trade-off.

Weaknesses:

  • Unaddressed computational overhead from TrueSkill (see detailed questions below).
  • Closely related paper needs to be added to related work.

问题

Minor / Typos

  • Line 80: missing space “TDPart[25]”
  • Line 149: “the the”
  • Line 189: was \tau defined?

局限性

See Weaknesses/Questions above.

最终评判理由

The authors adequtely addressed my concerns by providing detailed data showing the negligible computational overhead of TrueSkill and agreeing to add a missing related work. I would like to keep my recommendation (Accept).

格式问题

None

作者回复

We thank the reviewer for the positive and supportive comments. We are glad the reviewer found our probabilistic modeling approach effective and appreciated the reproducibility of our results. Minor concerns raised by the reviewer are addressed below.

1. Computational Overhead of TrueSkill

  • We thank the reviewer for raising this point. We measured and reported the computational overhead of TrueSkill at Table 7 (below), which was negligible compared to the cost of LLM reranker calls. On the DL19 benchmark, the total runtime of AcuRank was approximately 58.28 seconds, among which the cumulative time spent on TrueSkill updates was only 0.06 seconds (0.1%). The portion of running Trueskill versus the total process was consistent (0.1%) across subsets of data (DL19, DL20, TREC-COVID, News) and for AcuRank-9 and AcuRank. This was measured using time.time() around the call to trueskill.rate() (line 170 of run.py in the supplementary material).
  • The trueskill.rate() function performs lightweight updates to the mean and variance of document scores based on closed-form Gaussian updates (see Supplementary Section 2.3), which are efficiently computed on CPU. This confirms that nearly all computation time in AcuRank is spent on LLM calls, with minimal overhead from the uncertainty modeling.

Table 7: Comparing per-query end-to-end Latency with the TrueSkill overhead

LatencyDL19 (AcuRank‑9)DL19 (AcuRank)DL20 (AcuRank‑9)DL20 (AcuRank)TREC‑COVID (AcuRank‑9)TREC‑COVID (AcuRank)News (AcuRank‑9)News (AcuRank)
Total (s)58.28105.9859.2893.3067.32142.0883.65149.58
TrueSkill (s)0.060.110.060.100.060.130.060.11
TrueSkill ratio (%)0.10 %0.10 %0.10 %0.11 %0.09 %0.09 %0.07 %0.08 %

2. Usage and Computation of t(k)t(k)

  • We thank the reviewer for the clarification request. The threshold t(k)t(k) is indeed used in our algorithm. As described in Section 3.3, it is part of the Uncertainty-based Selection step (Line 3 of Algorithm 1), where it helps identify documents whose inclusion in the top-kk set is uncertain.
  • The formal definition of t(r)t(r) is provided in L. 154–157. It is the threshold for which the expected number of documents exceeding it in relevance equals rr, i.e., iP(xi>t(r))=r\sum_i \mathbb{P}(x_i > t(r)) = r. Each document's relevance xix_i is modeled as a Gaussian distribution determined by its current estimated mean μi\mu_i and variance σi2\sigma_i^2, so the value of t(r)t(r) depends on these parameters across all documents. While this equation does not admit a closed-form solution, we exploit its monotonicity to efficiently compute t(k)t(k) via binary search. We will clarify this explanation in the final version.

3. Related Work: Budget-Constrained Bayesian Pairwise Reranking

  • We thank the reviewer for bringing this work to our attention. To the best of our knowledge, the paper was made public shortly before our submission, and was under review and not yet publicly visible during the time of our research, so we were unable to cite the paper.
  • While both methods adopt TrueSkill-based uncertainty modeling, the objectives and settings differ significantly. That work uses a fixed-budget pairwise reranking strategy, whereas our approach performs listwise reranking with adaptive computation, dynamically allocating effort based on ranking uncertainty. This enables AcuRank to achieve both high efficiency and accuracy in zero-shot reranking scenarios.
  • Nontheless, since the paper is relevant to our work, we will include a discussion in the related work section of the final version.

4. Minor Issues

  • We thank the reviewer for pointing out the minor issues. We will correct the typos on Line 80 and Line 149. As for τ\tau in Line 189, we define it as the threshold at which the number of uncertain documents C|C| satisfies the stopping criterion. We will make this definition more explicit in the final version.
评论

Thanks for the clarifications.

审稿意见
4

This paper proposes a novel adaptive reranking framework called AcuRank, aiming to improve the efficiency and accuracy of listwise reranking with large language models (LLMs). The core idea is to dynamically adjust computational resources (including computational effort and objectives) based on uncertainty estimates of document relevance. AcuRank employs a Bayesian TrueSkill model to iteratively refine relevance estimates until a sufficient confidence level is reached. This explicit modeling of ranking uncertainty enables principled control over the reranking behavior, avoiding unnecessary computations on already confident predictions. The authors conducted experiments on TREC-DL and BEIR benchmarks, across different retrieval tasks and LLM-based reranking models. The results demonstrate that AcuRank consistently outperforms fixed-computation baselines, exhibiting excellent accuracy-efficiency trade-offs and better scalability with computational resources.

优缺点分析

Strengths:

  1. Clear Motivation Statement: The introduction clearly articulates the limitations of existing fixed-computation reranking methods and the motivation behind adopting an adaptive approach, presenting a very logical flow.

  2. Novelty and Significance: The concept of uncertainty-aware adaptive computation for LLM reranking is highly novel, directly addressing critical practical challenges such as the high inference cost and fixed context lengths of current LLM rerankers. Specifically, the use of the Bayesian TrueSkill model for explicitly modeling uncertainty offers a theoretically rigorous approach.

  3. Strong Experimental Results: The consistent improvements in accuracy-efficiency trade-off and better computational scalability demonstrated on established benchmarks (TREC-DL and BEIR), and with different LLM rerankers (RankZephyr, RankGPT), provide strong empirical support for AcuRank's effectiveness and generality.

Weaknesses:

  1. Insufficient Theoretical Depth: Complexity of Bayesian TrueSkill Model Integration: While the Bayesian TrueSkill model excels in game and competitive rankings, its application to the iterative process of document reranking, particularly the mapping of its inherent mechanisms and parameters (e.g., skill values, uncertainty, definition of "match outcomes"), is not immediately obvious. The paper could provide a deeper explanation of this part, for instance, how to precisely define the "match" results of documents during each reranking iteration, and how these results update the skill (relevance scores) and uncertainty in the TrueSkill model.

  2. Insufficient Theoretical/Empirical Basis for Uncertainty Thresholds: A core aspect of AcuRank lies in deciding when to stop based on uncertainty thresholds. However, the paper lacks sufficient theoretical guidance or sensitivity analysis regarding the selection of these thresholds (e.g., ϵstop and σtarget).

  3. Lack of Generalization Analysis across Different LLM Behaviors: The paper uses RankZephyr and RankGPT as LLM rerankers and shows AcuRank's effectiveness for both. However, there is a lack of additional baselines; for example, considering the impact of first-stage retrieval models like monoBERT, or comparisons with models like GPT-4/GPT-3.5.

  4. Unreasonable Evaluation of Efficiency. This paper uses the “number of reranker calls” to evaluate model efficiency. However, this is not an appropriate metric, as the time cost depends not only on the number of reranker calls but also on the context length of each call.

问题

Basis for ϵ stop and σ target Selection and Sensitivity Analysis: I noticed that AcuRank's performance heavily depends on the setting of uncertainty thresholds. Could the authors provide the rationale for choosing these hyperparameters and a more detailed sensitivity analysis?

局限性

Yes

最终评判理由

The new explanations and supplementary experiments have successfully resolved the key questions we raised. We now have a much better understanding of the work and are satisfied with the improvements. Based on these changes, we have updated our score for this paper from 2 to 4.

格式问题

No

作者回复

Thank you for providing valuable comments to improve our paper! We are especially encouraged by the recognition of AcuRank’s originality and its excellent trade-off between accuracy and efficiency. We hope our comments are sufficient enough to answer the questions raised and that you consider raising the overall rating of our work. We would greatly appreciate it if you could review our comments and inform us if any further discussions or clarification is required:

1. Justification for TrueSkill-based Uncertainty Estimation

  • We appreciate the reviewer’s request for a deeper explanation of how the TrueSkill model is applied in our setting and why it is suitable for estimating uncertainty in document relevance. As we provide a detailed justification in our response to Reviewer KzUm #3, we kindly refer the reviewer to that section for further discussion.

2. Choice of Uncertainty Thresholds

  • We thank the reviewer for raising this important point. In AcuRank, the uncertainty thresholds ϵ\epsilon and σ\sigma are adjustable parameters that control the trade-off between efficiency and accuracy. A relaxed setting (i.e., smaller ϵ\epsilon or σ\sigma) generally improves accuracy at the cost of increased computation, while stricter values reduce LLM usage with some performance degradation. This flexibility allows users to adapt the method to their specific resource and performance requirements.
  • Our configurations were chosen to reflect different trade-off points depending on user preference: AcuRank-H and AcuRank-HH favor accuracy, AcuRank-9 prioritizes efficiency, and the standard AcuRank strikes a practical balance.
  • In response to the reviewer’s suggestion, we report additional ablation results to illustrate how performance (in terms of NDCG@10) and computation vary with these thresholds. These results show that the trade-off curve is smooth and monotonic, suggesting that AcuRank is robust across a wide range of hyperparameter choices.

Table 3. Analysis on sensitivity to uncertainty thresholds (BM25 top-100, RankZephyr).

(ϵ,τ)(\epsilon, \tau)DL19DL20DL21DL22DL23DL-HCOVIDNFCSignalNewsR04ToucheDBPScifAvg.# Calls
(0.01,20)(0.01, 20)74.670.570.151.846.039.784.836.932.354.256.637.745.575.255.416.0
(0.01,10)(0.01, 10): AcuRank74.271.870.352.047.039.485.337.231.853.956.636.546.075.355.519.2
(0.01,5)(0.01, 5)73.871.871.051.847.339.885.137.332.254.057.335.345.875.155.528.2
(0.0001,10)(0.0001, 10): AcuRank-H74.670.870.552.247.340.485.837.432.153.756.837.546.075.455.741.7
(0.0001,5)(0.0001, 5): AcuRank-HH74.771.870.651.947.040.086.137.531.354.457.836.146.075.455.857.2

3. Generalization across Different LLM Rerankers and Retrieval Setups

  • We thank the reviewer for raising this important point. One of the key advantages of AcuRank is its general applicability: it can be applied after any first-stage retriever and on any listwise reranker, regardless of the underlying architecture or scoring function.
  • In addition to RankZephyr (main Table 1 & 2) and RankGPT (main Table 2), we also evaluate AcuRank with a diverse set of listwise rerankers, including RankVicuna (Supple Table 5), and Llama-3.3-70B-Instruct (Supple Table 6) to demonstrate its generalization across different LLM backbones. Publicly available listwise reranking models are relatively scarce, and we believe these models span a sufficiently diverse range of model types and capabilities. While additional results with GPT-4/3.5 could offer further insight, their high API cost made them infeasible for inclusion in our experiments.
  • We also evaluate AcuRank under various first-stage retrieval settings: BM25 top-100, BM25 top-1000, and SPLADE++ED top-100 (main Table 1 & 2). To consider the impact of first-stage retrieval models, as the reviewer requested, we additionally evaluate AcuRank on BEIR using contriever (facebook/contriever-msmarco, mean pooling and Pyserini indexing) as a first-stage retriever. As shown in Table 4 below, AcuRank again achieves superior accuracy-efficiency trade-off over baselines.
  • We will include these results on the final revision, and will open-source all related code and configurations to support reproducibility.

Table 4: Results with first-stage retrieval using Contriever top-100

MethodCOVIDNFCSignalNewsR04ToucheDBPScifAvg.# Calls
Contriever48.032.023.835.337.721.733.065.133.10
SW‑170.838.326.549.954.930.446.175.445.39
SW‑272.338.726.850.455.031.046.875.345.818
SW‑372.638.626.650.555.031.346.975.445.927
TourRank‑167.536.825.849.354.628.245.572.244.013
TourRank‑570.938.126.452.257.630.147.074.446.065
AcuRank74.439.127.153.057.233.147.975.247.420.1
AcuRank‑H74.339.227.652.458.032.148.076.447.443.4

4. Using the Number of Reranker Calls as an Efficiency Metric

  • We thank the reviewer for raising this point. The number of reranker calls was originally chosen as the efficiency metric since it is a practical proxy for end-to-end latency in real-world deployments (L. 209-213).
  • In fact, the number of reranker calls disadvantages AcuRank over baselines. Since AcuRank partitions uncertain candidates with maximum size of C(=20)C(=20), this leads to shorter inputs with smaller context numbers in some stages (L. 174-179). Therefore, AcuRank tends to get shorter inputs and runs faster compared to the fxed-computation baselines with similar number of reranker calls. Despite this potential disadvantage, AcuRank consistently demonstrates better efficiency.
  • To further support this point, we compare AcuRank and sliding window baselines (SW-1 and SW-2) using sampled queries with a similar number of reranker calls, and report both query-wise window size, input length, FLOPs, and end-to-end runtime latency across four sampled datasets (DL19, DL20, TREC-COVID, and News) at Table 5 and 6. These results confirm that AcuRank achieves comparable or better efficiency given similar number of reranker calls. Thank you for your suggestion, and we will additionally include the evaluation on our next revision.
  • In addition, as discussed in Supplementary L. 205–209, AcuRank can support parallel processing across disjoint candidate groups. This enables further latency reductions in practice beyond what is captured by reranker call counts alone.

Table 5: Additional evaluation of average per‑query efficiency — SW-2 vs. AcuRank

MetricDL19 (SW‑2)DL19 (AcuRank)DL20 (SW‑2)DL20 (AcuRank)TREC‑COVID (SW‑3)TREC‑COVID (AcuRank)News (SW‑2)News (AcuRank)
NDCG@1074.674.270.271.884.485.352.753.9
# Calls18.018.218.016.327.021.818.018.5
Window size20.016.620.016.720.016.320.016.2
Input length21561831216218223991375739923989
Latency (s)12610612693213142179150
petaFLOPs0.610.520.620.461.81.31.21.2

Table 6: Additional evaluation of average per‑query efficiency — SW-1 vs. AcuRank-9

MetricDL19 (SW‑1)DL19 (AcuRank‑9)DL20 (SW‑1)DL20 (AcuRank‑9)TREC‑COVID (SW‑1)TREC‑COVID (AcuRank‑9)News (SW‑1)News (AcuRank‑9)
NDCG@1074.073.370.271.484.183.252.353.0
# Calls9.09.09.08.99.09.09.09.0
Window size20.018.920.018.920.018.820.018.9
Input length21492058215220623990386539923987
Latency (s)9358625973678884
petaFLOPs0.300.290.310.290.590.570.590.59
评论

Dear Reviewer SkZp,

Thank you for your time and effort in reviewing our paper. We noticed that all of the other reviewers have responded, expressing that their concerns are addressed, and are satisfied with the additional results provided, and recommended acceptance. May we kindly ask if your concerns are addressed as well? As the discussion period is reaching to an end, we would greatly appreciate hearing your thoughts to ensure we have met your expectations. Should further clarification or additional experiments be helpful, we will be happy to provide them promptly. Thank you again for your guidance and for considering our work.

审稿意见
5

This paper proposes AcuRank, an uncertainty-based adaptive computation framework for listwise re-ranking. The main contributions are as follows:

  1. Introduction of a TrueSkill-based probabilistic relevance modeling approach, which enables the estimation of uncertainty in document ranking.
  2. Development of a novel listwise re-ranking framework that supports adaptive computation by iteratively refining and selectively re-ranking uncertain candidate documents.
  3. Extensive experimental validation across various benchmarks and datasets, demonstrating AcuRank’s superior trade-off between accuracy and efficiency, as well as its generalization capability in different retrieval tasks and LLM-based re-ranking models.

AcuRank employs a Bayesian TrueSkill model to maintain probabilistic estimates of document relevance, iteratively refining these estimates until a sufficient confidence level is reached. By explicitly modeling ranking uncertainty, this approach allows for rational control of re-ranking actions and avoids unnecessary updates for predictions that are already highly certain. On the TREC-DL and BEIR benchmarks, AcuRank consistently achieves a better balance between accuracy and efficiency compared to fixed-computation baselines.

优缺点分析

Strengths:

  1. AcuRank is the first to introduce uncertainty modeling into listwise re-ranking tasks, enabling fine-grained control over uncertainty through dynamic allocation of computational resources, which demonstrates high originality.
  2. Across multiple benchmarks, AcuRank achieves a better trade-off between accuracy and efficiency. For example, on the TREC-DL and BEIR datasets, AcuRank consistently improves ranking quality under various re-ranking models while requiring fewer re-ranker invocations.
  3. The paper is well-written and easy to understand.

Weaknesses:

The main limitation lies in insufficient experimental evaluation. The authors mention using TourRank and TrueSkill-Static as baselines; however, Table 1 does not report results for TrueSkill-Static, and the lower part of Table 1 lacks results for TourRank. Table 2 exhibits similar issues. To robustly assess AcuRank’s performance across retrieval scenarios, comprehensive comparisons with these baseline models are necessary. Additionally, a comparison with setwise methods is also needed.

问题

Q1: AcuRank dynamically allocates computational resources to handle uncertain documents. However, in certain scenarios, could this dynamic adjustment lead to either over-allocation or under-allocation of resources? For example, when processing highly complex queries, is there a risk that the model may terminate computation prematurely and consequently overlook some important documents?

Q2: AcuRank employs the TrueSkill model to estimate the uncertainty of document relevance. However, is this uncertainty estimation method sufficiently accurate?

局限性

yes

最终评判理由

All my concerns have been addressed. I recommend acceptance.

格式问题

No

作者回复

We thank the reviewer for the encouraging and constructive feedback. We appreciate the recognition of the clarity and practicality of our framework. We would greatly appreciate it if you could review our comments and inform us if any further discussion or clarification is required.

1. Missing Baseline Results

  • While it would be ideal to include all baseline results in every table, space limitations and computational costs led us to prioritize the most competitive and relevant baselines. Less impactful or redundant baselines were omitted to maintain clarity and focus. We include additional results below and are open to reporting others during the rebuttal period if time and resources allow.
  • TrueSkill-Static results on all datasets are already available in Supplementary Table 3. These were excluded from the main table due to space constraints and consistently underperform compared to AcuRank.
  • We additionally report TourRank baseline results missing from Table 1 and 2. We stopped at the point where TourRank-xx became significantly worse than AcuRank in terms of both the number of reranker calls and NDCG@10. Specifically, we include TourRank-1 in the lower part of Table 1 (BM25 top-1000), and both TourRank-1 and TourRank-2 in Table 2 (SPLADE++ED top-100). Thank you for pointing this out. We will incorporate these results in the final version.

Table 1: Additional baselines (TourRank-1) on the lower part of main paper Table 1 (BM25 top-1000, RankZephyr)

MethodDL19DL20DL21DL22DL23DL-HCOVIDNFCSignalNewsR04ToucheDBPScifAvg#pass
SW-175.178.871.557.549.540.980.738.028.951.048.030.948.076.456.294.6
TourRank-175.476.671.756.649.842.182.736.629.950.959.933.045.572.856.0117.1
AcuRank76.775.373.159.353.541.085.037.030.756.563.236.248.876.058.068.4

Table 2: Additional baselines (TourRank-1,2) on the upper part of main paper Table 2 (SPLADE top-100, RankZephyr)

MethodCOVIDNFCSignalNewsR04ToucheDBPScifAvg.# Calls
SW-185.237.529.750.162.028.949.375.752.39.0
TourRank-182.537.428.551.560.729.748.973.751.613
TourRank-284.637.629.552.561.728.450.074.052.326
AcuRank85.537.727.251.963.130.250.075.652.78.9
  • We additionally report results on another different first-stage retriever (contriever) which also confirms that AcuRank's superior accuracy-efficiency trade-off with varying first-stage retrievals. Please refer to our response to Reviewer SkZp for details on these results (Table 3).
  • In the related work section, we categorize reranking approaches into pointwise, pairwise, setwise, and listwise paradigms, and our focus in this paper is specifically on listwise reranking. Setwise methods are typically compared against pairwise methods and not compared with listwise methods, so we considered them less relevant to our approach. Nonetheless, we appreciate the suggestion and acknowledge their potential interest to some readers. We will include a discussion and comparison in the camera-ready version.

2. Possibility of Over- or Under-allocation of Computation

  • Since the ideal amount of computation for each case cannot be known in advance, adaptive computation may not guarantee optimal allocation of resources. However, AcuRank includes safeguards to reduce the risk of both over- or under-allocation. It adaptively allocates reranker calls based on estimated ranking uncertainty, focusing on documents whose inclusion in the top ranks remains uncertain, and stops when the system becomes sufficiently confident or reaches a predefined maximum number of iterations.
  • We also support our claim with two empirical analyses:
    • In Section 5.3 (Figure 3) and Supplementary Section 5.1, we show there is a positive correlation between weighted information gain (WIG, accounting for complex, confusing candidate document scenarios) and the number of reranker calls (Lines 306-315).
    • In Supplementary Section 5.2 (Table 8), we show that AcuRank tends to allocate more computation to harder queries (lower NDCG than the average) and less computation to easier ones, further validating the effectiveness of our strategy.
  • We also provide a qualitative example of the bayesian update at the response for reviewer nvvL (Number 2, Table 9,10).

3. Justification for TrueSkill-based Uncertainty Estimation

  • TrueSkill is a principled Bayesian rating system originally developed for competitive games, where each player's skill is modeled as a Gaussian distribution that is updated through observed match outcomes. It is known to produce well-calibrated estimates of of both scores and uncertainty, and has been widely adopted in various applications where reliable uncertainty quantification is crucial [1,2,3].
  • Similar to prior works, we recast documents as players and interpret the LLM reranker’s output rankings as pairwise comparisons. When one document is ranked above another within a batch, it is treated as having won that match. TrueSkill uses these outcomes to iteratively update each document’s relevance score (mean) and uncertainty (variance), enabling it to estimate both how relevant each document is and how confident it is in those estimates.
  • Although the accuracy of uncertainty estimates cannot be directly evaluated without ground-truths, the strong accuracy-efficiency trade-offs achieved by AcuRank demonstrate that the uncertainty signals derived from TrueSkill effectively guide adaptive computation.

[1] Chung et al., CHI 2022, TaleBrush: Sketching Stories with Generative Pretrained Language Models

[2] Park et al., UIST 2023, Generative Agents: Interactive Simulacra of Human Behavior

[3] Shaikh et al., CHI 2024, Rehearsal: Simulating Conflict to Teach Conflict Resolution

评论

Thank you to the authors for your response. My questions have been resolved. Overall, this is an excellent paper, and I maintain my decision to accept it.

最终决定

This work addresses the problem of listwise ranking of documents with LLMs. Specifically, it highlights a key limitation of listwise rankers that attempt to balance ranking effectiveness with computational cost. It claims that listwise rankers are inflexible as they do not adapt to query complexity/ambiguity and assign fixed number of LLM reranking calls to predetermined number of candidates irrespective of complexity/ambiguity. It proposes a remedy to this limitation by adaptively allocating LLM reranking calls to candidates based on the candidate's uncertainty relative to the top-k boundary. Towards this end, it maintains a probabilistic relevance estimate of each candidate using a Bayesian rating system and iteratively updating the estimates based on the observed reranking results. Until convergence, it works with only the uncertain candidates in each iteration while skipping the more certain candidates and thereby allocating LLM reranking call to only the most uncertain candidates. It claims that this adaptive uncertainty-aware allocation of computation and reranking improves both accuracy and computational efficiency. The proposed approach is tested on TREC Deep Learning and BEIR benchmarks showing improvement in ranking quality while simultaneously requiring fewer LLM reranking calls compared to the chosen baseline reranking algorithms.

Unlike prior work in document relevance that use point estimates for the relevance of candidates, the current work models the relevance of each candidate as a Gaussian variable whose mean is the estimated relevance and variance captures both the epistemic uncertainty and global noise in the estimation. While the global noise factor is independent of the candidate and remains constant throughout, the epistemic uncertainty factor is expected to go down iteratively as the model becomes more confident about the relevance of the candidate. This is an interesting and potentially useful approach for document reranking but is a straightforward adaptation of a well-known prior work, TrueSkill to document retrieval task. The key innovation in the current work is to employ the document relevance distributions to derive useful signals for adaptive resource allocation. Specifically, the probability that a document has rank r and therefore is in the top-k can be computed. However, as this computation is expensive, a computationally more efficient approximation is used by defining a threshold t(r) such that the expected number of candidates whose relevance exceeds this threshold equals the target rank r and then computing the probability that the current candidate's relevance exceeds the threshold.

The algorithm for reranking with adaptive computation works in iterations. In each iteration, the algorithm selects a subset of candidates based on their ranking uncertainty. To compute a candidate's uncertainty, the algorithm computes the probability that its relevance exceeds a threshold for rank. If this probability is within a range determined by a hyperparameter, the candidate is regarded as uncertain. The subset of candidates selected in this manner are then partitioned into equal sized groups based on their relevance. Candidates with similar relevance are highly likely to fall in the same group and it is argued that reranking candidates with similar relevance scores yields higher information gain. The algorithm passes each group of candidates to LLM for reranking and uses the resulting ranking to update TrueSkill parameters. The distributions of candidates whose ranking improved/worsened substantially relative to their relevance based ranking are updated substantially. The iterative refining of the distributions continues till any of the stopping criteria is met. The algorithm produces the final ranking of the candidates according to the decreasing order of their final relevance estimate.

Experimental study is centered around TREC-DL and BEIR benchmarks, retrieval effectiveness is reported using NDCG@10 metric and efficiency is measured as number of reranker calls per query. BM25-Top100, BM25-Top100 and SPLADE++ED-Top100 are used as the first stage retrievers. RankZephyr and RankGPT are used as the rerankers with a candidate list size of 20. Sliding windows, TourRank and TrueSkill-Static are used as the listwise ranking baselines.

At the lowest #Calls setting reported (Table 1), the proposed algorithm achieves the same or better NDCG@10 on 8 datasets and worse NDCG@10 on 6 datasets compared to the strongest performing baseline SW-1. The average gain in NDCG@10 over all the datasets is 0.3 relative to SW-1. As #Calls increases, the algorithm demonstrates better retrieval effectiveness than the baselines but the improvement is incremental. Ablative study demonstrates the effect of the design choices (TruSkill Score Initialization, Partitioning and Stopping Criteria) on effectiveness and efficiency.

The paper is well-motivated and well-written. The idea of adaptively allocating reranker calls to documents based on their uncertainty is interesting and novel. Use of TrueSkill to iteratively refine candidate relevance is sensible though straightforward. Experimental study is designed reasonably well to demonstrate the benefits of the proposed algorithm over the baselines. Improvement in retrieval effectiveness over the baselines is incremental though not consistent across the datasets. Authors were responsive to the concerns and questions fielded by the reviewers in the review as well as rebuttal discussions and the concerns were largely addressed. Overall, the work presents an interesting approach to adaptive computation with promising results.