PaperHub
6.3
/10
Poster3 位审稿人
最低3最高4标准差0.5
3
4
3
ICML 2025

Reward-Guided Speculative Decoding for Efficient LLM Reasoning

OpenReviewPDF
提交: 2025-01-22更新: 2025-07-24
TL;DR

Reward-Guided Speculative Decoding for faster inference and better accuracy.

摘要

关键词
Large language modelsProcess reward modelSpeculative decoding

评审与讨论

审稿意见
3

The paper introduces Reward-Guided Speculative Decoding (RSD), an improved version of SD which replies on a process reward model to determine the quality of a step instead of exact match. RSD combines a lightweight draft model with a more capable target model, integrating a reward-based mechanism to optimize computational cost and output quality. RSD allows controlled bias by accepting high-reward outputs from the draft model, even when they do not perfectly match the target model’s predictions. The authors provide theoretical justification for a threshold-based mixture strategy, ensuring an optimal balance between computational efficiency and performance. Empirical results on various reasoning benchmarks, including GSM8K, MATH500, OlympiadBench, GPQA, MMLU STEM, and GaoKao-2023-En, demonstrate that RSD achieves up to 4.4× fewer FLOPs while improving accuracy by up to +3.5 points compared to standard speculative decoding.

给作者的问题

No.

论据与证据

Most claims in the paper are well-supported with both theoretical analysis and empirical evidence:

  1. The claim that RSD reduces computational cost while maintaining or improving accuracy is supported through detailed FLOPs analysis and comparisons across multiple benchmarks.
  2. The claim that RSD outperforms traditional speculative decoding (SD) on reasoning tasks is substantiated with controlled experiments using various LLM sizes and evaluation datasets.

Some aspects could benefit from additional justification:

  1. The assertion that RSD can “even surpass the large model’s performance” requires more scrutiny. While some results suggest that RSD achieves higher accuracy than the single target model, this may depend on careful hyperparameter tuning (e.g., choice of threshold δ), and the exact conditions under which this occurs need further clarification.
  2. The theoretical analysis of reward-guided acceptance could be extended by considering alternative reward models or potential biases introduced by specific reward function choices.

Overall, the claims are convincing, but further ablation studies on reward function variations and distribution mismatch effects would strengthen the argument.

方法与评估标准

Yes, the proposed methods and evaluation criteria are well-aligned with the problem of efficient inference for LLM reasoning:

  1. Evaluation on a broad set of reasoning benchmarks (GSM8K, MATH500, OlympiadBench, etc.) is appropriate because speculative decoding has traditionally been underutilized for multi-step reasoning tasks.
  2. Comparisons with baseline methods (single target model, speculative decoding, Best-of-N, beam search) ensure a fair assessment of RSD's performance.
  3. Computational efficiency is rigorously analyzed using FLOPs per question, which is a relevant metric for assessing inference cost.

However, a more detailed breakdown of performance by question difficulty (e.g., comparing RSD with SD for simple vs. complex problems) could provide deeper insights into when RSD is most beneficial.

The chosen evaluation methods are sound, but the paper could further explore RSD’s robustness to different reward models and variations in reasoning complexity.

理论论述

The proofs in the paper are logically consistent, they assume an idealized reward function without considering potential noise in reward estimation. A discussion on the sensitivity of these theoretical results to imperfect reward models would be useful.

实验设计与分析

I reviewed the experimental setup, and it appears to be methodologically sound:

  1. Controlled comparisons with speculative decoding and single-model baselines ensure a fair evaluation.
  2. Hyperparameter tuning for δ is systematically explored, showing that a threshold around δ = 0.7 balances efficiency and accuracy.
  3. Computational cost is measured in FLOPs, providing an objective comparison.

Limits:

  1. One possible limitation is that the experiments primarily focus on math and reasoning tasks. It would be interesting to see how RSD performs on open-ended generation tasks (e.g., summarization, dialogue).
  2. The choice of process reward model (PRM) is not extensively analyzed—would a different PRM significantly change RSD’s effectiveness?

Overall, the experimental setup is robust, but future work could explore how different PRMs influence performance and whether RSD generalizes beyond reasoning tasks.

补充材料

The proof is pretty extensive in the Appendix B, and more results in Appendix C.

与现有文献的关系

The paper builds on prior work in speculative decoding and efficient inference for LLMs, making several novel contributions:

  1. Improves upon speculative decoding (SD): Prior SD methods (Leviathan et al., 2023) enforce strict unbiasedness, whereas RSD introduces reward-guided acceptance to enhance efficiency.
  2. Incorporates a process reward model (PRM): Similar to reinforcement learning approaches (Dong et al., 2023), but applied to speculative decoding for stepwise validation of reasoning.
  3. Demonstrates efficiency gains over standard decoding strategies: The results align with recent efforts to reduce inference cost via speculative sampling (Chen et al., 2023) and model acceleration (Frantar et al., 2022).

遗漏的重要参考文献

No

其他优缺点

Weakness: The performance of the PRM is quite important for the process, yet this is can also be the limit of performance. Although the result shows it can go beyond the target model, but still limited.

其他意见或建议

The authors discussed about the choice of \delta and some experiments to explore the pick of it. I am wondering if it is possible to make the selection of the parameter as dynamic based on the difficulty of the question.

作者回复

 ~

Supplementary Material: https://anonymous.4open.science/r/Rebuttal-supp-6595-C4A2/appendix_rebuttal_ICML.pdf

 ~


Q1. Why RSD outperforms the large model?

Please refer to our rebuttal to Reviewer p3vx Q5.

 ~

Q2. Alternative reward models or potential biases

As we mentioned, one might consider alternative formulations:

  • Likelihood-Ratio-Based: One could incorporate a term like PM(yz)Pm(yz)\frac{\mathbf{P}_M(y|z)}{\mathbf{P}_m(y|z)} to define the weighting function (e.g., w(yz)=min(1,αPM(yz)Pm(yz))w(y|z)=\min(1, \alpha\,\frac{\mathbf{P}_M(y|z)}{\mathbf{P}_m(y|z)})). Such a formulation may capture discrepancies between the two models’ distributions, but it may also introduce bias if the likelihood estimates are themselves noisy or miscalibrated.
  • Hybrid: One might combine the process reward and likelihood ratio—e.g., w(yz)=min(1,βr(yz)PM(yz)Pm(yz))w(y|z)=\min(1, \beta\,r(y|z)\,\frac{\mathbf{P}_M(y|z)}{\mathbf{P}_m(y|z)}). While this hybrid approach can potentially leverage complementary strengths, its sensitivity to misestimations in either component may induce systematic biases. For example, if the reward function overestimates quality in regions where the draft model already performs well, the weighting may overfavor the draft model, reducing the benefits of the target model’s corrections.

In both cases, under our assumption, the underlying distribution will be better than small model induced one. Whereas, the choice of reward function can significantly affect the acceptance probability, and thus the final mixture distribution. When we choose more weights towards likelihood, it will biased to exact-match of large model distribution, so that the cost would be higher (similar to SD).

You may also consider imperfect reward (Q5).

 ~

Q3. Performance by question difficulty.

We show it in Tab C.4 (in link). One can observe:

  • For simpler questions (level=1), RSD and SD performs the same;
  • For harder questions (level>2), RSD consistently outperforms SD. With the help of PRM, RSD selectively includes the target model for some reasoning steps, correcting the wrong reasoning step and leading to better performance.

 ~

Q4. Different reward models & variations in reasoning complexity.

In Tab C.5 (in link), we includes more PRMs, Qwen2.5-Math-PRM-7B and Qwen2.5-Math-PRM-72B, released during the reviewing cycle.

  • RSD is robust to PRM. Although Skywork-o1-Open-PRM (regression model) and Qwen2.5-Math-PRM (classification model with num_labels=2) are trained differently, they all perform quite well across different tasks.
  • δ=0.7\delta=0.7 consistently yields strong results across all four PRMs, appearing to be a sweet spot for balancing efficiency and performance when the reward score lies in [0, 1].

We also draw the accuracy and efficiency wrt δ\delta in Fig C.3 (in link). These two PRMs behave similarly:

  • Involving the target model (δ0\delta \neq 0) consistently improves accuracy over using the draft model alone (δ=0\delta = 0), with greater gains on harder questions.
  • For the same δ\delta, the proportion of questions solved by the draft model alone decreases with an increasing level, showing that harder questions need more involvement of the target model.

 ~

Q5. Imperfect reward models

In practice, r^(yz)\hat{r}(y|z) is an approximation of the oracle reward roracle(yz)r_{\text{oracle}}(y|z). We can derive non-asymptotic convergence bounds under assumptions on the reward model’s error distribution.

Assumption (Sub-Gaussian Error): Assume that for each reasoning step, the estimation error

ϵ(yz)=r^(yz)roracle(yz)\epsilon(y|z) = \hat{r}(y|z) - r_{\text{oracle}}(y|z)

is sub-Gaussian with parameter σ2\sigma^2; that is, for any t>0t > 0,

Pr(ϵ(yz)t)2exp(t22σ2).\Pr\left(|\epsilon(y|z)| \geq t\right) \leq 2\exp\left(-\frac{t^2}{2\sigma^2}\right).

Under this assumption, when using nn independent observations, standard concentration inequalities imply that, with probability at least 1δ1-\delta,

1ni=1nr^(yiz)E[roracle(yz)]σ2log(2/δ)n.\left|\frac{1}{n}\sum_{i=1}^n \hat{r}(y_i|z) - \mathbf{E}[r_{\text{oracle}}(y|z)]\right| \leq \sigma \sqrt{\frac{2\log(2/\delta)}{n}}.

Even if the reward model is imperfect, the empirical average reward converges to the oracle at O(log(1/δ)n)O\Bigl(\sqrt{\frac{\log(1/\delta)}{n}}\Bigr). The bias introduced in the acceptance decision and in PRSD\mathbf{P}_{\text{RSD}} can be controlled in a non-asymptotic manner. Even with imperfect reward estimates, the impact on the RSD diminishes as more data become available.

 ~

Q6. Open-ended generation tasks.

Refer to Q2 of Reviewer GXeJ.

 ~

Q7. Choice of δ\delta based on the difficulty.

From Fig C.1, δ=0.7\delta=0.7 performs consistently well across difficulty levels, achieving top accuracy in 4/5 levels. Q4 further confirms δ=0.7\delta=0.7 works well for different PRMs.

That said, dynamically selecting δ\delta based on question difficulty (eg using a LLM to predict difficulty and adjust δ\delta accordingly) could improve performance and efficiency. We leave this for future work.

审稿意见
4

This paper introduces Reward-Guided Speculative Decoding (RSD), a novel framework designed to improve the efficiency of inference in large language models (LLMs) by combining a lightweight draft model with a more powerful target model. Extensive evaluations on challenging reasoning benchmarks demonstrate that RSD significantly reduces computational costs (up to 4.4× fewer FLOPs) while improving accuracy compared to using the target model alone or parallel decoding methods. The results highlight RSD as a robust and cost-effective approach for deploying LLMs in resource-intensive scenarios, particularly for complex reasoning tasks.

给作者的问题

Could the framework be extended to handle multimodal inputs?

论据与证据

The paper claims that Reward-Guided Speculative Decoding (RSD) improves LLM inference efficiency by combining a draft model with a target model, guided by a process reward model. Evidence includes up to 4.4× fewer FLOPs and improved accuracy on reasoning benchmarks like MATH500 and GSM8K. The results show RSD outperforms traditional methods and models.

方法与评估标准

RSD uses a draft model to generate candidate outputs, which are evaluated by a reward model to decide whether to invoke the target model. Evaluation is based on computational efficiency (FLOPs) and accuracy across reasoning tasks. Benchmarks include GSM8K, MATH500, and Olympiad-level tasks.

理论论述

The paper claims that a threshold-based mixture strategy optimally balances computational cost and output quality. Theoretical analysis shows that RSD ensures higher expected rewards compared to using the draft model alone.

实验设计与分析

Experiments compare RSD to baseline methods like speculative decoding and target-only models across multiple reasoning benchmarks. Metrics include FLOPs, accuracy, and the proportion of questions solved by the draft model alone. Ablation studies explore the impact of threshold values and weighting functions.

补充材料

The appendix includes proofs for theoretical propositions, additional experimental results, and details on model merging. It also provides a discussion on the relationship between RSD and speculative decoding, as well as the impact of different weighting functions.

与现有文献的关系

RSD builds on speculative decoding and process reward models, addressing limitations in traditional methods. It aligns with research on efficient LLM inference and reasoning tasks.

遗漏的重要参考文献

N/A

其他优缺点

A weakness is the reliance on a process reward model, which may introduce additional overhead.

其他意见或建议

How does RSD perform on non-reasoning tasks like text generation or summarization?

作者回复

Q1. The reliance on a process reward model, which may introduce additional overhead.

This is indeed a very practical concern, since RSD utilizes one more model, i.e. the process reward model (PRM), than speculative decoding (SD). However, accoding to our experiments, the overhead is minor from three perspectives:

  • The PRM is small: According to Table 2, a 1.5B PRM could already offers consistent better performance than all baselines with different draft and target models.
  • The inference cost of PRM is minimal: RSD applies PRM to score the step instead of tokens from the draft model. The average number of steps per question in MATH500 is 18. The inference cost of PRM is similar to generate 18 tokens per question.
  • The PRM can be merged with the target or draft model: In Table 4, we explore the possibility of merging the PRM with the draft or target model. Surprisingly, model merging doesn't obviously degrade the performance, and even results in better performance when merging the larger models. In this way, RSD shares the same number of models as SD. A further investigation of model merging between PRM and proxy model is left to future work.

 ~

Q2. How does RSD perform on non-reasoning tasks like text generation or summarization?

This is a very interesting suggestion. An important component of RSD is PRM. As far as we know, there is not yet a PRM for general-domain generation. But there are plenty of outcome reward models (ORMs) for open-ended generation. Could we use ORM instead of PRM in RSD?

Experimental setup: We utilize Llama-3.2-1B-Instruct as the draft model, Llama-3.1-8B-Instruct as the target model, and Skywork-Reward-Llama-3.1-8B-v0.2 as the ORM. The 805 prompts from AlpacaEval [1] are used for the generation, and the model outputs are evaluated with AlpacaEval2.0 against the outputs from gpt4_turbo. Similar to the setting in the paper, we define a generation ended with "\n\n" as a reasoning step, and apply the ORM to score this step. The score of Skywork-Reward-Llama-3.1-8B-v0.2 ranges from -\infty to \infty. We didn't extensively tune the reward threshold δ\delta, and empirically chose δ=0\delta=0.

Results: As shown in the following table, even with an ORM instead of a PRM, RSD achieves a significantly better win rate than the draft model, showing RSD's robustness across different tasks. Among all generated tokens, 65% tokens are generated by the draft model only without any intervention of the target model. We believe that a general-domain PRM and dedicated tuning of δ\delta could further boost the performance.

MethodWin Rate (%) against gpt4_turbo
Single Draft Model7.09
Single Target Model24.47
RSD18.85

 ~

Q3. Could the framework be extended to handle multimodal inputs?

Yes, we believe our RSD could be seamlessly extended to the multomodal reasoning tasks, because:

  • The key component of RSD is PRM. As shown in Q2, even an ORM could act as a PRM. We believe that the existing multimodal ORM can be used in RSD for multimodal reasoning tasks.
  • We noticed that a new multimodal PRM [2] is released recently. It can be used perfectly in our framework.

However, the exploration of multimodal reasoning tasks is out of the scope of this work, and there isn't any multimodal PRM when we conducted this research, we leave this investigation to interested readers.

 ~

[1] Length-Controlled AlpacaEval: A Simple Way to Debias Automatic Evaluators. Yann Dubois, Balázs Galambosi, Percy Liang, Tatsunori B. Hashimoto

[2] VisualPRM: An Effective Process Reward Model for Multimodal Reasoning. Weiyun Wang, Zhangwei Gao, et. al., Wenhai Wang

 ~


Thank you very much for your thoughtful suggestions, making our work more solid. We have incorporated the new results in the updated version by our side.

If these revisions address your concerns, we kindly request a reconsideration of the scores. Should you have any further questions, we are happy to assist.

审稿人评论

Thanks for the detailed reply. My concerns have been resolved. I increased the rate for this paper.

作者评论

Dear Reviewer GXeJ,

We are very encouraged by your increased rate.

We really enjoy the discussion, and thank you for your suggestion about open-ended generation with RSD.

We believe this suggestion makes our work more solid and strong.

Best!

审稿意见
3

This paper introduces a method to guide speculative decoding using a reward model. Unlike standard speculative decoding, where a larger model verifies the outputs of a smaller model, the proposed approach determines acceptance or rejection based on reward signals. Specifically, the authors design a series of weighting functions to normalize the reward as a probability that stochastically decides whether to accept or reject the generated responses. Experimental results demonstrate that this method maintains efficiency while improving reasoning performance.

给作者的问题

The ablation results indicate that the threshold δ\delta plays a crucial role in reasoning accuracy but is highly sensitive across different tasks. Given its definition in Proposition 2.3, can δ\delta be estimated beforehand based on theoretical results, rather than relying solely on empirical tuning?

论据与证据

Most of the claims in this paper are clear and well-supported. However, the statement that “High-quality tokens (e.g., those favored by a process reward) may still be rejected if their probabilities under the large model are too low” raises some concerns. This claim suggests a potential misalignment between the large model and the reward model. However, the proposed algorithm still trusts the large model’s output to correct responses rejected by the reward model (as shown in Eq. 1). This implies that while the large model has high precision, it may also exhibit a high false positive rate. Further justification is needed on this to explain why the reward model serves as a more reliable verifier than the large model’s likelihood or a combination of both.

方法与评估标准

The proposed methods are evaluated on 6 math reasoning benchmarks, which align with the major claim and the choice of reward models. The experiments also include a comparison on FLOPs.

理论论述

The theoretical results are partially checked (Props 2.1,2.2,2.3). It remains unclear what the “largest possible threshold that makes the function satisfy the constraint” refers to until I reach the appendix.

实验设计与分析

The experimental designs are overall sound. More details on how FLOPs are recommended to be included. It is also suggested to include more efficiency metrics, such as throughput (tokens per secs).

补充材料

The proofs were checked. The remaining parts were roughly gone through.

与现有文献的关系

The proposed method has relevance to recent topics scaling inference-time compute to facilitate LLM reasoning, in which an LLM samples the solution according to a reward to do reasoning.

遗漏的重要参考文献

No critical references known to me are missing.

其他优缺点

Strength:

  • The overall writing quality is good, with clear mathematical formulations that effectively convey intuition to the reader. I found the results and discussions in Appendix C.2 particularly insightful, as they illustrate how RSD dynamically allocates computational resources based on task complexity.
  • The proposed approach is simple yet effective, offering a novel way to leverage reward models not only for enhancing reasoning accuracy but also for improving efficiency.
  • The experimental results are compelling, demonstrating that reward-guided decoding can even surpass the performance of the strongest large model.

Weakness:

  • The theoretical results do not fully explain why RSD outperforms the large model.
  • The motivation for why the reward model serves as a better acceptance criterion than the larger model is not entirely convincing. Since the algorithm ultimately relies on the larger model for corrections, it raises the question of why the reward model improves the verification process over large models’ likelihood. See “Claims And Evidence” for more details.
  • Additionally, while speculative decoding in RSD is guided by process rewards, the baselines rely on outcome rewards. To ensure a fair comparison, it would be beneficial to include additional search-based baselines that also incorporate process rewards.

其他意见或建议

Ln 30-31, “significant better” -> “significantly better”. Ln 325-326, “majotity” -> “majority”

作者回复

Q1. PRM vs large models’ likelihood

The core issue is the misalignment between the large model (LM) and PRM, leading SD to reject high-quality tokens due to low LM likelihood. While the reviewer suggests this implies high FP rates from LM, these are often style-/format-related, not correctness errors. LMs generalize better but are NOT trained for correctness. In contrast, PRMs are explicitly trained to identify correctness, making them more suitable for detecting errors, while LM likelihood would point out both style inconsistency and errors. Thus, our algorithm uses LM guidance only after PRM rejection, leveraging both PRM's specialized correctness identification and LM’s strength in generalizing better completions. Using both for rejection is also promising; we leave it for future work as noted in Sec 2.4.

 ~

Q2. Calculation of FLOPs.

As described in Sec 3.3, we follow the standard FLOPs approximation for transformer models with N parameters, i.e., approximately 2N FLOPs per inference token, as adopted in prior works [1,2]. We provide a detailed example below: The FLOPs calculation for RSD (7B/72B/7B) is presented in the table below.

TargetDraftPRM
Model Size72B7B7B
Tokens/Question6739618
TFLOPs9.6485.5440.252

[1] Scaling laws for neural language models.

[2] Beyond chinchilla-optimal: Accounting for inference in language model scaling laws.

 ~

Q3. More efficiency metrics (throughput).

We measure the througput and provide the following table. We use batch size 256 and MATH500. The observation is similar to Figure 4: RSD is faster than SD.

MethodThroughput
Single Draft (1.5B)1.00 ×\times (4697 tokens/s)
Single Target (7B)0.70 ×\times
SD(1.5B/7B)0.83 ×\times
RSD(1.5B/7B/1.5B)0.91 ×\times
RSD(1.5B/7B/7B)0.85 ×\times

 ~

Q4. Search-based baselines

We compare RSD to the search-based baselines (Process Best-of-N and Beam Search) in Table 3, where RSD significantly outperforms the search-based baselines that only utilize a draft model and PRM. It shows the importance of involving a larger model in the reasoning path for correction.

 ~

Q5. Why RSD outperforms the large model?

Proposition (Improved Expected Reward via RSD): Assume that for each decoding step the reward function r(yz)[0,1]r(y|z)\in[0,1] and consider ω(r)=1(rδ)\omega(r) = \mathbf{1}(r \geq \delta) for some threshold δ[0,1]\delta \in [0,1]. Define the RSD mixture distribution as PRSD(yz)=ω(r(yz))Pm(yz)+νPM(yz),\mathbf{P} _ {\text{RSD}}(y| z) = \omega(r(y| z)) \mathbf{P} _ m(y | z) + \nu\mathbf{P} _ M(y| z), where ν=1PyPm{r(yz)δ}\nu = 1 - \mathbf{P} _ {y\sim \mathbf{P} _ m}\{r(y| z) \geq \delta\}. The expected reward of the RSD distribution is EPRSD[r(yz)]=αEPm[r(yz)r(yz)δ]+(1α)EPM[r(yz)],\mathbf{E} _ {\mathbf{P} _ {\text{RSD}}}[r(y|z)] = \alpha\,\mathbb{E} _ {\mathbf{P}_m}[r(y| z)| r(y| z) \geq \delta] + (1-\alpha)\,\mathbb{E} _ {\mathbf{P}_M}[r(y| z)], with α=PyPm{r(yz)δ}.\alpha = \mathbf{P} _ {y\sim \mathbf{P}_m}\{r(y| z) \geq \delta\}.

The RSD distribution outperforms the target model PM\mathbf{P} _ M (i.e., EPRSD[r(yz)]>EPM[r(yz)]\mathbf{E} _ {\mathbf{P} _ {\text{RSD}}}[r(y| z)] > \mathbf{E} _ {\mathbf{P} _ M}[r(y| z)]) if and only if EPm[r(yz)r(yz)δ]>EPM[r(yz)].\mathbf{E} _ {\mathbf{P}_m}[ r(y| z) | r(y| z) \geq \delta] > \mathbf{E} _ {P_M}[r(y| z)].

Thus, RSD yields a higher expected reward than PM\mathbf{P}_M provided that the subset of tokens generated by the draft model Pm\mathbf{P}_m (i.e. those with r(yz)δr(y| z) \geq \delta) has a higher reward than PM\mathbf{P}_M. If one can choose δ\delta so that only high-quality tokens from the draft model are accepted, then the mixture of these with the fallback tokens from PM\mathbf{P}_M lifts the overall expected reward of RSD above that of the large model alone.

This result offers a theoretical explanation for why RSD can outperform the large model: it leverages the fact that, when the draft model’s top-performing outputs (as measured by rr) are even better on average than the target model’s outputs, selectively accepting these outputs increases the overall quality.

In summary, for RSD to be advantageous over PM\mathbf{P}_M, one must choose a threshold δ\delta such that: EPm[r(yz)r(yz)δ]>EPM[r(yz)].\mathbf{E} _ {\mathbf{P}_m}[r(y| z)| r(y| z) \geq \delta] > \mathbf{E} _ {\mathbf{P}_M}[r(y| z)].

Under this condition, RSD leads to an improved expected reward compared to PM\mathbf{P}_M.

 ~

Q6. Estimation of δ

A6. As noted in Prop. 2.3, δ\delta corresponds to a quantile of the reward distribution under a given compute budget. While theoretical guidance informs its range, estimating the exact quantile requires access to empirical reward statistics -- hence, practical tuning (e.g., grid search) is necessary and suggested by our analysis. Moreover, Table 2 demonstrates that performance is not highly sensitive to δ\delta, suggesting our method is robust across a reasonable range and not reliant on fine-tuning.

最终决定

This paper introduces process reward models (PRMs) into the speculative decoding framework. Instead of requiring the large model to strictly verify the small model’s output, a reward model is used to judge whether the small model’s output is acceptable. This allows token-level differences, reducing the intervention frequency from the large model. As a result, inference becomes more efficient without sacrificing reasoning quality. The method achieves up to 4.4× speedup compared to using the large model alone, and improves accuracy by up to 3.5 points over parallel decoding with the small model on reasoning tasks.

All reviewers are in favor of acceptance. The paper presents well-designed experiments and introduces insightful ideas. I recommend acceptance.

One minor suggestion, raised by reviewers p3vx and XUtc, concerns the criterion for selecting the reward threshold δ\delta. The paper would benefit from a clearer explanation of how δ\delta is chosen, along with an analysis of its impact across broader domains to support practical applications.