PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
4
5
5
4
3.8
置信度
创新性3.0
质量2.8
清晰度2.5
重要性3.0
NeurIPS 2025

A*-Thought: Efficient Reasoning via Bidirectional Compression for Low-Resource Settings

OpenReviewPDF
提交: 2025-05-11更新: 2025-10-29

摘要

Large Reasoning Models (LRMs) achieve superior performance by extending the thought length. However, a lengthy thinking trajectory leads to reduced efficiency. Most of the existing methods are stuck in the assumption of overthinking and attempt to reason efficiently by compressing the Chain-of-Thought, but this often leads to performance degradation. To address this problem, we introduce A*-Thought, an efficient tree search-based unified framework designed to identify and isolate the most essential thoughts from the extensive reasoning chains produced by these models. It formulates the reasoning process of LRMs as a search tree, where each node represents a reasoning span in the giant reasoning space. By combining the A* search algorithm with a cost function specific to the reasoning path, it can efficiently compress the chain of thought and determine a reasoning path with high information density and low cost. In addition, we also propose a bidirectional importance estimation mechanism, which further refines this search process and enhances its efficiency beyond uniform sampling. Extensive experiments on several advanced math tasks show that A*-Thought effectively balances performance and efficiency over a huge search space. Specifically, A*-Thought can improve the performance of QwQ-32B by 2.39$\times$ with low-budget and reduce the length of the output token by nearly 50% with high-budget. The proposed method is also compatible with several other LRMs, demonstrating its generalization capability.
关键词
long-to-shortthought compressionreasoningefficient CoT

评审与讨论

审稿意见
4

The paper introduces A*-Thought, a novel framework designed to enhance the reasoning efficiency of Large Reasoning Models (LRMs) by compressing their lengthy reasoning chains. The authors propose a bidirectional importance estimation mechanism to evaluate the significance of each reasoning step based on its relevance to both the question and the solution. This mechanism assigns a bidirectional importance score (BIS) to each step, which is then used in conjunction with the A* search algorithm to identify and isolate the most essential thoughts from the extensive reasoning chains. The A* search algorithm employs cost functions that assess the quality of the current path and estimate the future cost to reach a desirable final solution, thereby efficiently compressing the chain of thought and determining a reasoning path with high information density and low cost. Extensive experiments on various advanced math tasks demonstrate that A*-Thought effectively balances performance and efficiency over a huge search space. For instance, it can improve the performance of QwQ-32B by 2.39× with low-budget and reduce the length of the output token by nearly 50% with high-budget. The proposed method is also compatible with several other LRMs, showcasing its generalization capability.

优缺点分析

Strengths

  1. The paper is novel and presents a well-defined problem formulation. Both the Step-Level Bidirectional Importance Score and the Path-Level A* Search are intuitively motivated and thoroughly described.
  2. The proposed method achieves strong empirical performance, demonstrating a 2.39× speedup over the 32B model.
  3. The overall structure of the paper is clear and easy to follow.

Weaknesses

  1. It would be helpful to include a more detailed analysis of the computational cost incurred by the A* search process when selecting training samples across benchmarks.
  2. The paper would benefit from a broader evaluation across different model scales to better assess robustness.

问题

Questions

  1. What validation model is used during the A* search process? The paper does not seem to specify this, which raises concerns about reproducibility and implementation clarity.

  2. The teaser figure is difficult to interpret. What do the circles of different sizes and depths represent? A legend or further explanation would greatly improve readability.

  3. Does using \n\n as a delimiter for segmenting thought steps risk omitting essential intermediate reasoning steps, especially in multi-line or complex CoT outputs?

  4. Have the authors explored test-time scaling strategies, such as self-consistency or diverse sampling? Such evaluations would help better understand the upper-bound performance of the method.

局限性

No.

最终评判理由

All of my concerns have been addressed. I have updated my score based the rebuttal.

格式问题

No.

作者回复

For Question 1

The verification model V\mathcal{V} is the original s1.1-32B model (reported in the Training Details paragraph of Section 4.1, line 178).

For Question 2

We have already explained the meaning of the circles and their color intensity (reported in the description of Figure 1).

To be precise, all circles are of a consistent size, with each representing a thought step.

The color of each circle indicates its importance, a value quantified by the Bidirectional Importance Score (BIS) from Section 3.1. Consequently, a darker shade signifies a higher importance score.

For Question 3

Following prior works such as Retro-Search, ThinkPrune, Length-filtered Vote, and SEAL, we split reasoning steps using "\n\n".

Building on this, we are the first to propose a "thinking span" strategy. When an individual thought step t(n)\mathbf{t}^{(n)} is to be explored, we instead select a span centered around it:r(n)=t(n1),t(n),t(n+1),\mathbf{r}^{(n)}=\langle \mathbf{t}^{(n-1)}, \mathbf{t}^{(n)}, \mathbf{t}^{(n+1)}\rangle,where t(n1)\mathbf{t}^{(n-1)} and t(n+1)\mathbf{t}^{(n+1)} are the immediately preceding and succeeding steps of t(n)\mathbf{t}^{(n)}. This method allows the A* search process to simultaneously consider both the physical order of the original thought path and the logical, importance-based order of the BIS queue.

For Question 4

We are grateful to the reviewer for this insightful suggestion. We reran the A*-Thought evaluation on MATH-500 and LiveCodeBench, this time using a best-of-N strategy (with N=8) to calculate pass@k. The findings indicate that A*-Thought and test-time-scaling strategies can be used in combination. These results will be integrated into the final version of the paper.

MethodsMATH-500LiveCodeBench
pass@1pass@2pass@4pass@8pass@1pass@2pass@4pass@8
Budget: 512
QwQ-32B10.817.225.934.40.00.00.00.0
+ A*-Thought (Ours)33.247.460.770.24.57.012.520.8
Budget: 1K
QwQ-32B16.629.642.153.20.00.10.10.3
+ A*-Thought (Ours)50.865.475.881.811.820.431.944.8
Budget: 2K
QwQ-32B51.259.867.874.82.13.45.17.0
+ A*-Thought (Ours)69.279.285.489.024.534.045.256.5
Budget: 4K
QwQ-32B75.480.284.086.812.417.723.028.0
+ A*-Thought (Ours)78.886.891.193.239.050.960.968.8

For Weakness 1

Our training data is independent of the test sets (reported in the Training Data and Verification Model paragraph of Section 4.1, lines 177-178).

A*-Thought was run on the 1K training data, this involved a 5.2-minute BIS scoring stage and a 7.43-hour A* search stage. For different benchmarks, the computational cost of our method is then proportional to the length of the output tokens required at inference time.

Crucially, A*-Thought also demonstrates remarkable generalization. Though trained only on math, our supplementary evaluations on LiveCodeBench and MMLU confirm that it successfully applies the A*-Thought pattern to non-math tasks (e.g., coding, history, computer science, law).

MethodsLiveCodeBenchMMLUAverageACU
Acc.Len.Acc.Len.Acc.Len.
Budget: 512
QwQ-32B0.0512.0037.6511.9418.8511.973.67
+ A*-Thought4.5509.5356.7398.5730.6454.056.74
Budget: 1K
QwQ-32B0.01021.9457.4956.9028.7989.422.90
+ A*-Thought11.8986.0271.9573.3041.9779.665.37
Budget: 2K
QwQ-32B3.51977.5875.21323.2639.41650.422.38
+ A*-Thought24.51734.2879.0671.4151.81202.854.30
Budget: 4K
QwQ-32B12.03586.9379.51584.0545.82585.491.77
+ A*-Thought39.03044.5180.3733.9059.71889.213.16

For Weakness 2

We are grateful for the reviewer's insightful suggestion. We conducted new experiments by training the Qwen2.5-7B-Instruct and Qwen2.5-14B-Instruct on data compressed by A*-Thought and benchmarking their performance against the TokenSkip baseline. The results affirm that A*-Thought is robust and maintains its effectiveness across different model sizes. These new results will be incorporated into the final manuscript.

MethodsMATH500AMC23OlympiadBenchGSM8KAverageACU
Acc.Len.Acc.Len.Acc.Len.Acc.Len.Acc.Len.
Budget: 512
TokenSkip-7B9.4511.412.5512.004.0512.0025.5508.4010.4511.02.03
A*-Thought-7B (Ours)30.8499.125.0508.706.7509.3659.2447.2225.4491.105.18
TokenSkip-14B9.2510.897.5512.004.8512.0029.0511.1712.6511.522.47
A*-Thought-14B (Ours)30.4497.5410.0512.0010.4508.7750.9475.3025.4498.405.10
Budget: 1K
TokenSkip-7B24.8986.6210.01021.337.91019.1151.0907.4023.4983.622.38
A*-Thought-7B (Ours)48.2866.9325.0991.4219.3961.4274.3661.9941.7870.444.79
TokenSkip-14B23.6991.6510.01024.007.41020.2247.8960.1322.2999.002.22
A*-Thought-14B (Ours)49.2873.5322.5997.3318.8966.5578.3738.6142.2894.014.72
Budget: 2K
TokenSkip-7B35.41776.5020.01968.4714.71959.1355.61534.3731.41809.621.74
A*-Thought-7B (Ours)59.21352.5430.01715.2225.41677.3877.6874.1148.11404.813.42
TokenSkip-14B55.81678.0320.01934.0520.81940.5873.81453.7142.61751.592.43
A*-Thought-14B (Ours)64.21308.1840.01650.4028.31610.0486.1930.8954.71374.883.97
Budget: 4K
TokenSkip-7B39.83201.9027.53593.9320.93566.0063.42651.4037.93253.311.16
A*-Thought-7B (Ours)61.81999.1030.02840.7029.63048.7378.01271.1849.92289.932.18
TokenSkip-14B68.42406.1640.03338.8032.53334.7381.22054.7855.52783.621.99
A*-Thought-14B (Ours)70.81677.0042.52350.3234.62494.2488.81100.5259.21905.523.11
审稿意见
5

This paper presents an approach for decreasing the length of chain-of-thought reasoning traces while maintaining accuracy. To do so, they design an algorithm that searches for subtraces that can still reach the final answer, and then conduct supervised fine-tuning (I think?) on those subtraces. Evaluated on mathematical reasoning benchmarks, the resulting models are much more accurate for a given token budget, and maintain accuracy when answering without a budget too. The algorithm for discovering subtraces is said to be a form of A* search that sorts tokens based on their importance in determining the solution, then chooses which tokens to in an expand-then-greedily-select step until the current subtrace is verified to produce a correct solution.

优缺点分析

Strengths

The approach works really well compared to baselines. The paper is well-written. Most of it is easy to understand and clear. The approach is interesting and novel. At first when I read the paper, I was skeptical that such an approach would lead to useful results, so it's surprising to see that it works. The interesting part of the approach is the fine-grained importance score as the basis of discovering useful traces. It's based on attention weights and the likelihood of the model producing the solution given the tokens up to that point, but using GPT-2 instead of whatever model is generating the reasoning traces. I find it quite surprising that GPT-2 could yield useful scores this way, given how weak of a model it is.

Weaknesses

The A* algorithm and the bidirectional score could be explained more clearly. See questions section. There are some other confusing parts of the paper, see questions section. The baselines the approach is being compared to seem quite weak — see appendix E for prompts.

问题

Q1. Is a reasoning step t(n)t^(n) a single token? Or is it like a sentence? If it's not a token, is the splitting of an answer into thoughts given by your dataset? For the rest of my answer, I'll assume it's a single token.. Q1. For the NLL part of the BIS, it looks like NLL(st(n))NLL(s | t^(n)) is the probability of generating the solution given the single reasoning step? Isn't that going to be very hard for the model? Similarly, it seems odd to calculate the NLL(t(n)q)NLL(t^(n) | q). But I guess even though it's weird, there's enough signal there to be useful? Q3. I don't understand the AA^* search algorithm. Is this even A^*? Here is my understanding: 1. sort tokens by BIS score. start with empty trace T. 2. take the top W tokens. For each one, score using cost function f. Take the best scoring token and add it to the trace. 3. check if current trace leads to solution using verifier. If it does, return current trace. Otherwise, repeat step 2.

Questions about this algorithm: Q3.1: why are there two different scoring functions, BIS and the cost function? It seems like BIS more efficient to evaluate, and f more expensive? Why not just use the same exact function for the two, but use a cheap model for the initial sorting function and then the full model for the cost function? Q3.2: what is the value of W? is this stated somewhere? Q3.3: the tokens after being sorted by BIS are out of order, right? when added to the trace, they are put back into their order, right? Q3.4: after expanding W options, are the ones that aren't finally selected put back into the queue, or discarded? Q4. how does the verifier work? You just say which model it is, but I don't understand what the model is doing.

局限性

N/A

最终评判理由

The authors did a good job explaining details of the approach that I was confused about. I will update my score to a 5.

格式问题

N/A

作者回复

For Q1

Each thought step t(n)\mathbf{t}^{(n)} is a complete sentence (reported in Section 3.1, line 105).

Following prior works such as Retro-Search, ThinkPrune, Length-filtered Vote, and SEAL, we split reasoning steps using \n\n. Our process begins by extracting the content between the model-specific identifiers, <think> and </think>. This content is then segmented by \n\n to generate a set of thought-step sentences. Following the subsequent bidirectional importance scoring and A* search phases, we identify a minimal subset of these steps, which forms the compressed reasoning output.

For Q 2

NLL(yx)\mathrm{NLL}(\mathbf{y} | \mathbf{x}) reflects the probability of obtaining y\mathbf{y} under the condition of x\mathbf{x}, where x\mathbf{x} and y\mathbf{y} are distinct sentences.

Given that each thought step t(n)\mathbf{t}^{(n)} is a complete sentence, we can easily calculate Negative Log-Likelihood (NLL) using language models (e.g., GPT-2):

y_j | \mathbf{x}, \mathbf{y}_{<j} \right ).$$ A more informative context $\mathbf{x}$ increases the likelihood of generating the target $\mathbf{y}$. We apply this by treating each thought step $\mathbf{t}^{(n)}$  as the context $\mathbf{x}$ and calculating its relevance to both the question $\mathbf{q}$ and the solution $\mathbf{s}$ (as the target $\mathbf{y}$). This results in two key scores: $\mathrm{NLL}\left( \mathbf{q} | \mathbf{t}^{(n)} \right)$ and $\mathrm{NLL}\left( \mathbf{s} | \mathbf{t}^{(n)} \right)$. To improve scoring and enhance the distinctiveness of the results, we follow LongLLMLingua by inserting a template phrase to explicitly prompt the causal connection. Specifically, we use: - For $\mathrm{NLL}\left( \mathbf{q} | \mathbf{t}^{(n)} \right)$: "We can get the solution to this question in the given thought step." - For $\mathrm{NLL}\left( \mathbf{s} | \mathbf{t}^{(n)} \right)$: "We can get this solution in the given thought step." # For Q 3 A* search is a canonical heuristic algorithm in computer science, which was jointly proposed by Peter Hart, Nils Nilsson and Bertram Raphael in 1968. It is widely applied in the fields of path planning and map navigation in AI systems. It uniquely combines search efficiency with the guarantee of discovering the shortest possible path. The algorithm progresses by calculating a total cost function, $F(n)=G(n)+H(n)$, for every node n it considers. Within this function, $G(n)$ represents the exact cost incurred to travel from the origin to node n, while $H(n)$ represents the future cost—a heuristic estimation of the cost from node n to the final destination. It is by virtue of this heuristic information, $H(n)$, that the A* algorithm can intelligently navigate a vast search space without resorting to an exhaustive exploration, thus ensuring computational efficiency. Drawing an analogy, we can view the standard Chain-of-Thought (CoT) method, shown in Figure 1(a), as an unguided traversal of all intermediate nodes, which is often a convoluted and verbose approach. Inspired by the principles of A* search, our A*-Thought method (Figure 1(b)) introduces a heuristic to this process. Here, the importance of each node—visualized by its color depth—serves as the heuristic. By integrating this information, A*-Thought efficiently searches for the most concise thought trajectory required to solve a problem, which in turn leads to highly efficient reasoning and inference. Therefore, the overall pipeline is: 1. sort **`sentences`** by BIS score. start with empty trace T. 2. take the top W **`sentences`**, and respectively obtain the **`thinking span`** centered on the sentence and carrying the preceding and following sentences. For each one, score using cost function F. Take the best scoring **`thinking span`** and add it to the trace. 3. check if current trace leads to solution using verifier. If it does, return current trace. Otherwise, repeat step 2. # For Q 3.1 Both the Bidirectional Importance Score (BIS) and the Cost Function serve to guide the expansion of the search tree, but at different granularities: BIS determines the **coarse-grained global direction**, while the Cost Function determines the **fine-grained local direction**. - **Global Direction (BIS)**: In classic pathfinding, the positions of all nodes between the start and end points are fixed, which allows for a straightforward determination of candidate node priority at each search depth. In contrast, for the many thought steps in a Chain-of-Thought, we must first assess the intrinsic importance of each sentence based on its quality and informational content. This assessment establishes a priority order for expanding candidate steps into the search tree. Therefore, the role of BIS is to guide the global direction of the search. - **Local Direction (Cost Function):** At each step of the search, we examine only the top $W$ candidates from the BIS queue. We form a "thinking span" around each candidate and apply the cost function $f(\cdot)$ to select the single best span to advance the search path, returning the other candidates to the queue. This makes the Cost Function a local, step-by-step decision-maker. A key distinction lies in their nature: **`BIS scores are static`**, calculated once to reflect the intrinsic merit of each step. In contrast, the **`Cost Function is dynamic`**. Navigating a non-enumerable search space efficiently requires a dynamic function that can accurately assess a path's current cost and future value at every step. This conceptual separation informs our choice of models. For the global ranking task performed by BIS, we use the efficient and cost-effective GPT-2 model. For the critical, dynamic, and fine-grained local evaluation managed by the Cost Function, we leverage the more powerful and precise s1.1-32B model to achieve superior performance. # For Q 3.2 The exploration size $W$ was set to 2 (already reported in the Training Details paragraph of the original Section 4.1, line 206). # For Q 3.3 We use the Bidirectional Importance Score (BIS) to sort the sentences in **`descending order`**, forming a priority queue. In this queue, the sentences are intentionally out of sequence compared to their order in the original CoT. However, when they are selected and added to the compressed path, they are arranged to maintain their **`original order`**. # For Q 3.4 After expanding the $W$ options, candidates that are not selected are **`put back into the queue`**. # For Q 4 Given the current compressed thinking path $\mathbf{t}^{\prime}_k$, question $\mathbf{q}$ and solution $\mathbf{s}$, the validation model $\mathcal{V}$ plays a dual role during the A* search phase: **generating a candidate solution $\mathbf{s}^{\prime}_k$ for verification and calculating the Cost Function $f(\cdot)$ for exploration** (reported in Section 3.1, lines 132-139 and 158-167). ## 1. Verification Phase First, the validation model $\mathcal{V}$, is prompted to perform sequence completion on the following input: $$\mathbf{q} \langle think \rangle \mathbf{t}_k^{\prime} \langle/think\rangle \langle solution \rangle$$ This generates a candidate solution $\mathbf{s}^{\prime}_k$, resulting in the complete sequence: $$\mathbf{q} \langle think \rangle \mathbf{t}^{\prime}_k \langle /think \rangle \langle solution \rangle \mathbf{s}^{\prime}_k \langle /solution \rangle$$ If the generated solution matches the ground truth ($\mathbf{s}^{\prime}_k==\mathbf{s}$), it indicates that the current path $\mathbf{t}^{\prime}_k$​ is of high quality and contains sufficient information. The algorithm then terminates and returns $\mathbf{t}^{\prime}_k$​ as the final output. If they do not match ($\mathbf{s}^{\prime}_k\neq\mathbf{s}$), the algorithm proceeds to the exploration phase. ## 2. Exploration Phase In this phase, the top $W$ "thinking spans" from the priority queue, denoted as $\lbrace \mathbf{r}_1, \dots, \mathbf{r}_W \rbrace$, are selected as candidates for expansion. For each candidate $\mathbf{r}_w$, the validation model $\mathcal{V}$ is used to evaluate the cost function $f\left (\mathbf{t}_k^{\prime}+\mathbf{r}_w\right) = g\left (\mathbf{t}_k^{\prime}+\mathbf{r}_w\right) + h\left (\mathbf{t}_k^{\prime}+\mathbf{r}_w\right)$. This is done by providing the model with the following input, which includes the ground-truth solution: $$\mathbf{q} \langle think \rangle \mathbf{t}^{\prime}_k+\mathbf{r}_w \langle /think \rangle \langle solution \rangle \mathbf{s} \langle /solution \rangle$$ The candidate with the minimum cost function value is then chosen to expand the search tree, and the process moves to the next iteration. Further details on the algorithm and the validation model's operation are provided in Algorithm 1 in Appendix A, lines 382-383. # For Weakness To comprehensively evaluate its performance, in addition to the prompt-based baselines (Chain-of-Draft and Break-the-Chain in Appendix E), we also compared A*-Thought against a training-based baseline and the backbone model. - **Training-based baseline:** We compared against **`TokenSkip`**, a method that involves a two-stage process: first, shortening long CoT datasets with LLMLingua-2 compression, and second, using the resulting data to train an efficient model. - **Backbone model baseline:** We also included the QwQ-32B model as a fundamental point of comparison.
评论

Thank you for your response and for addressing my confusion on many of the technical details.

审稿意见
5

The paper proposes A*-Thought, an inference-time framework that prunes large language models’ long chain-of-thought (CoT) traces. Each step receives a bidirectional importance score (BIS) derived from attention statistics and token-level NLL with respect to both the question and the predicted answer. An A* search then selects a compact subset of spans that minimises a cost combining current-path verification loss and a heuristic future cost. Experiments on four math-reasoning benchmarks with three 32 B models show accuracy gains (up to +18 pp) and shorter outputs (≈50 % fewer tokens) under fixed token budgets compared with prompt- and training-based baselines.

优缺点分析

Strengths

  • Potential to lower inference cost of LRMs, an important problem.

  • Empirical improvements on math tasks under tight budgets (Table 1).

Weaknesses

  • Evaluation limited to synthetic math datasets; no evidence on open-domain, multimodal or non-numeric reasoning.

  • Variance/error bars omitted despite stochastic search; single-run results may be fragile.

  • Conceptually close to Retro-Search and TokenSkip; contribution may be incremental (reuse of A* and attention/NLL saliency).

  • Gains are shown only for already-compressed 32 B distillations; unclear impact on frontier ( >70 B) models or real latency on hardware.

  • Training still demands multi-epoch passes over the entire CoT corpus (8 × A100 80 G, 3 epochs) — energy savings may be overstated.

问题

Q1: Have you tried GSM-like natural-language tasks, code reasoning, or commonsense QA to demonstrate domain robustness?

Q2: The verifier V is distilled from the same dataset used for pruning; could this inflate acceptance probability?

While the idea is interesting and results promising, the narrow evaluation scope, missing robustness analyses, and incomplete limitation discussion prevent a confident acceptance at this stage.

局限性

They do not discuss: (a) sensitivity to α, β, k_max across tasks; (b) potential semantic drift when critical but low-BIS steps are pruned; (c) ethical risks of deploying compressed reasoning that may omit safety checks. Hence the limitations section is insufficient.

最终评判理由

A*-Thought presents an interesting approach. The paper is well-written, demonstrating promising performance across datasets in different domains and achieving SOTA results.

格式问题

N/A.

作者回复

For Question 1 / Weakness 1

We thank the reviewer for their valuable suggestion. We tested A*-Thought's generalization on the out-domain benchmarks LiveCodeBench and MMLU across diverse domains such as coding, history, computer science, law.

The results demonstrate that the model successfully learned the A*-Thought pattern, achieving higher efficiency and performance on non-mathematical tasks. We will incorporate these findings into the final version of the paper.

MethodsMMLULiveCodeBenchAverageACU
Acc.Len.Acc.Len.Acc.Len.
Budget: 512
QwQ-32B37.6511.940.0512.0018.8511.973.67
+ A*-Thought56.7398.574.5509.5330.6454.056.74
Budget: 1K
QwQ-32B57.4956.900.01021.9428.7989.422.90
+ A*-Thought71.9573.3011.8986.0241.9779.665.37
Budget: 2K
QwQ-32B75.21323.263.51977.5839.41650.422.38
+ A*-Thought79.0671.4124.51734.2851.81202.854.30
Budget: 4K
QwQ-32B79.51584.0512.03586.9345.82585.491.77
+ A*-Thought80.3733.9039.03044.5159.71889.213.16

For Question 2

Verification model V\mathcal{V} is the original model of s1.1-32B, which was fine-tuned on the original s1K-1.1 dataset to preserve its long reasoning abilities. It is worth noting that the verification model was not fine-tuned on the compressed data.

For Limitation (a)

We are grateful for the reviewer's insightful suggestion. Accordingly, we have performed a sensitivity analysis to explore the impact of varying the hyperparameters  α\alpha, β\beta, kmaxk_{\mathrm{max}}. This was done by evaluating the A*-Thought-QwQ-32B model on the ARC and LiveCodeBench benchmarks with 512 tokens budget. We will update the final version of our paper with these experimental results.

MethodsARCLiveCodeBenchAverageACU
Acc.Len.Acc.Len.Acc.Len.
α=0.0\alpha=0.065.9405.315.5509.3135.70457.317.81
α=0.5\alpha=0.563.5381.024.5509.5334.00445.287.64
α=1.0\alpha=1.065.0393.208.5505.5936.75449.408.18
β=0.1\beta=0.163.5381.024.5509.5334.00445.287.64
β=0.5\beta=0.552.5438.034.0510.7328.25474.385.96
β=0.9\beta=0.948.2469.225.8506.1527.00487.695.54
kmax=10k_{\mathrm{max}}=1035.3476.902.2506.6818.75491.793.81
kmax=15k_{\mathrm{max}}=1554.2436.413.8510.2529.00473.336.13
kmax=20k_{\mathrm{max}}=2063.5381.024.5509.5334.00445.287.64

For Limitation (b)

Theoretically, Bidirectional Importance Score (BIS) is designed to find the most important reasoning steps. However, some key steps, like formulas, may get low scores when they're isolated.

To address this, we propose the "thinking span" strategy. Instead of picking just one step t(n)\mathbf{t}^{(n)} in A* search, we grab it along with its immediate neighbors: r(n)=t(n1),t(n),t(n+1),\mathbf{r}^{(n)}=\langle \mathbf{t}^{(n-1)}, \mathbf{t}^{(n)}, \mathbf{t}^{(n+1)}\rangle, where t(n1)\mathbf{t}^{(n-1)} and t(n+1)\mathbf{t}^{(n+1)} are the immediately preceding and succeeding steps of t(n)\mathbf{t}^{(n)}. This method preserves local context, counteracting the effects of information fragmentation and reducing potential semantic drift.

For Limitation (c)

Since the validation model is fine-tuned from the safety-aligned Qwen2.5-32B-Instruct, it also provides a degree of filtering for potentially harmful content during the A* search process.

For Weakness 2

We thank the reviewer for their valuable suggestion. To make the experimental results more robust, we conducted multiple-runs and reasoned 5 times, taking the average as the final result, and attaching the overall standard deviation after the result.

MethodsMATH500AMC23OlympiadBenchGSM8KAverageACU
Acc.Len.Acc.Len.Acc.Len.Acc.Len.Acc.Len.
Budget: 512
QwQ-32B10.8±0.710.8\pm0.7512.00±0.0512.00\pm0.02.5±0.02.5\pm0.0512.00±0.0512.00\pm0.03.8±0.33.8\pm0.3512.00±0.0512.00\pm0.028.7±1.428.7\pm1.4511.90±0.1511.90\pm0.111.5±0.511.5\pm0.5511.98±0.0511.98\pm0.02.24±0.12.24\pm0.1
+ A*-Thought33.5±0.933.5\pm0.9492.46±1.0492.46\pm1.018.5±2.018.5\pm2.0509.00±2.9509.00\pm2.911.4±0.611.4\pm0.6509.00±0.7509.00\pm0.758.3±0.758.3\pm0.7453.84±2.0453.84\pm2.030.4±0.730.4\pm0.7491.08±0.5491.08\pm0.56.19±0.16.19\pm0.1
Budget: 1K
QwQ-32B17.0±0.917.0\pm0.91018.39±1.11018.39\pm1.111.0±4.911.0\pm4.91024.00±0.01024.00\pm0.06.4±0.26.4\pm0.21023.97±0.01023.97\pm0.050.0±1.250.0\pm1.2950.39±2.4950.39\pm2.421.1±1.521.1\pm1.51004.19±0.41004.19\pm0.42.10±0.22.10\pm0.2
+ A*-Thought50.3±2.150.3\pm2.1853.22±4.3853.22\pm4.331.5±4.931.5\pm4.9944.83±14.7944.83\pm14.720.3±1.120.3\pm1.1962.95±6.3962.95\pm6.381.6±0.581.6\pm0.5696.85±4.6696.85\pm4.645.9±1.845.9\pm1.8864.47±5.1864.47\pm5.15.31±0.25.31\pm0.2
Budget: 2K
QwQ-32B50.6±1.150.6\pm1.11845.24±3.21845.24\pm3.226.0±1.226.0\pm1.21970.87±9.51970.87\pm9.518.0±1.018.0\pm1.02020.94±2.32020.94\pm2.380.3±0.680.3\pm0.61239.78±4.01239.78\pm4.043.7±0.543.7\pm0.51769.21±3.91769.21\pm3.92.47±0.02.47\pm0.0
+ A*-Thought69.2±0.869.2\pm0.81282.09±30.41282.09\pm30.444.0±5.644.0\pm5.61509.43±40.01509.43\pm40.031.1±1.131.1\pm1.11640.21±8.01640.21\pm8.091.6±0.491.6\pm0.4835.68±5.6835.68\pm5.659.0±1.359.0\pm1.31316.85±12.81316.85\pm12.84.48±0.14.48\pm0.1
Budget: 4K
QwQ-32B75.3±0.475.3\pm0.42776.35±15.62776.35\pm15.650.0±6.350.0\pm6.33485.04±39.63485.04\pm39.635.8±0.835.8\pm0.83634.92±9.43634.92\pm9.484.8±0.684.8\pm0.61362.99±8.71362.99\pm8.761.5±1.661.5\pm1.62814.83±11.82814.83\pm11.82.18±0.12.18\pm0.1
+ A*-Thought77.8±1.777.8\pm1.71717.41±17.11717.41\pm17.160.0±4.560.0\pm4.52286.07±49.92286.07\pm49.941.4±1.941.4\pm1.92535.87±28.02535.87\pm28.093.4±0.393.4\pm0.3883.02±17.6883.02\pm17.668.1±1.668.1\pm1.61855.59±12.51855.59\pm12.53.67±0.13.67\pm0.1

For Weakness 3

To our knowledge, A*-Thought is the first work to adapt the principles of A* search for this task.

Our method is fundamentally different from these predecessors.

  • vs. Retro-Search: A*-Thought performs a systematic, global search based on importance. Retro-Search does a local, sequential exploration.
  • vs. TokenSkip: Our BIS metric is both question and solution aware, whereas TokenSkip's is agnostic to global context information. We compress at the sentence level to preserve meaning, while TokenSkip works at the token level. Thus, despite a shared objective, our methodology is unique.

Beyond just guiding A* search, BIS metric also serves as a novel metric for analyzing the intrinsic importance mechanisms within long CoT. This aligns with a growing body of recent work (e.g., Token Entropy, Thought Anchors). We hope it provides a solid foundation for future research in this area.

For Weakness 4

Following DeepSeek-R1-Distill-Qwen, s1, SkyThought, Bespoke-Stratos, 32B size is strong mid-scale model that is sufficient for long CoT. Based on this, we have centered our efforts on the 32B model scale.

For Weakness 5

On our 8 A100 GPUs, training the s1-32B on its original 1K dataset took 3.84 hours, training Bespoke-Stratos-32B on its 17K dataset required 27 hours.

In contrast, after using A*-Thought to compress the 1K-row dataset, reducing the average data length by 68.69%. We trained our model for 3 epochs in 2.91 hours. reducing the training time by 24.25%, demonstrating a significant decrease in computational cost.

评论

A*-Thought has demonstrated SOTA performance across multiple domains. After reviewing the comments from other reviewers, I find that all my concerns have been addressed. I believe this paper is acceptable and thus am happy to raise my score to 5 (Accept).

审稿意见
4

This paper introduces A-Thought*, a framework that compresses lengthy Chain-of-Thought (CoT) outputs from large reasoning models using an A* search guided by a bidirectional importance score (BIS). The method addresses the efficiency challenges posed by verbose CoT reasoning, aiming to retain performance while significantly reducing token output.

优缺点分析

Strengths

  1. Clear Motivation and Relevance: The problem of excessive reasoning length in LRMs is well-known, and addressing it aligns with growing interest in cost-efficient inference.

  2. Methodological Innovation: The integration of a bidirectional importance score with A* search to identify high-value reasoning steps is novel.

  3. Empirical Performance: Substantial gains are reported in accuracy and accuracy per computation unit (ACU), especially under constrained budgets (e.g., 2.39× improvement for QwQ-32B).

Weaknesses

  1. Lack of Theoretical Analysis: There is no rigorous theoretical justification for why BIS + A* would yield better CoT compression than alternatives.

  2. Limited Performance Gains in High-Budget Scenarios: While A*-Thought shows impressive improvements under strict token constraints, its advantages diminish—and even reverse—when larger token budgets are available. In high-budget settings (e.g., 4096 tokens), A*-Thought underperforms compared to several baselines in terms of raw accuracy. This suggests that the proposed method may overly prioritize brevity at the cost of completeness.

问题

  1. Can you explain why A*-Thought underperforms in high-budget scenarios? Is there an inherent trade-off in the compression strategy that degrades solution completeness?

  2. How sensitive is the method to the selection of α\alpha (in BIS) and β\beta (in the cost function)? Have you performed any systematic hyperparameter tuning?

  3. Have you tested A*-Thought on non-mathematical reasoning tasks (e.g., commonsense QA, multi-hop QA)? If not, what are the anticipated limitations?

局限性

Yes

最终评判理由

Thanks for the authors' detailed response. I appreciate the results from the response and believe that it would make the paper more solid if the final version could include these results. However, I believe that my initial evaluation is accurate, as this paper is upper the borderline, therefore I decided to keep my original score.

格式问题

NA

作者回复

For Question 1 / Weakness 2

Based on the supplementary experiment (kmin=15k_{\mathrm{min}}=15) and the original experiment (kmin=5k_{\mathrm{min}}=5), it can be concluded that: The trade-off is that a smaller kmink_{\mathrm{min}} is more effective for low-budget reasoning, while a larger kmink_{\mathrm{min}} is more effective for high-budget reasoning.

Due to the historical difficulties associated with achieving lossless compression at the text level, the trade-off between performance and efficiency is generally regarded as inevitable. The A*-Thought algorithm employs a minimum search depth,  kmink_{\mathrm{min}} , to mitigate the negative semantic impact of overly short thought sequences. Verification of a thought path only occurs when its depth, kk, meets the condition  kkmink \ge k_{\mathrm{min}}.

However, using an excessively low kmink_{\mathrm{min}} value will result in the overall length of CoT data being too short, leading to performance degradation in high-budget scenarios. To address this, we explored increasing the value of kmink_{\mathrm{min}}​. The experiment results are shown in the following table. Notably, with kmin=15k_{\mathrm{min}}=15 in 4K budget, the A*-Thought model surpassed the performance of the strongest baseline, QwQ-32B fine-tuned with s1K-1.1, across all benchmarks. Furthermore, this best result was achieved with fewer inference tokens. We will update the final version of our paper with these findings.

MethodsMATH500AMC23OlympiadBenchGSM8KAverageACU
Budget: 4KAcc.Len.Acc.Len.Acc.Len.Acc.Len.Acc.Len.
QwQ-32B75.42798.6755.03456.0536.53645.2285.81348.2463.22812.052.25
QwQ-32B w/ s1K-1.179.62693.2765.03485.9542.43500.6695.21624.1170.62826.002.50
+ A*-Thought kmin=5k_{\mathrm{min}}=578.81699.7865.02385.8540.12546.4593.1874.5469.31876.663.69
+ A*-Thought kmin=15k_{\mathrm{min}}=1580.82184.3467.52893.6844.23063.9695.51229.4072.02342.853.07

For Question 2

For α\alpha, which is used to balance between question and solution information within BIS, systematic hyperparameter tuning has been reported in Figure 9 of Appendix D, varying α\alpha across the discrete set of values [0, 0.25, 0.5, 0.75, 1]. Experiments show that the best model performance can be achieved when BIS effectively integrates problems and solutions.

For β\beta, it is used to adjust the weight of the current cost function g()g(\cdot) in the overall cost function f()f(\cdot). In the following supplementary experiments, we discussed its discrete values in [0.1, 0.5, 0.9] on the AI2 Reasoning Challenge (ARC) and LiveCodeBench. We will add these experimental results to the final version of the paper.

MethodsARCLiveCodeBenchAverageACU
Acc.Len.Acc.Len.Acc.Len.
β=0.1\beta=0.163.5381.024.5509.5334.00445.287.64
β=0.5\beta=0.552.5438.034.0510.7328.25474.385.96
β=0.9\beta=0.948.2469.225.8506.1527.00487.695.54

For Question 3

We thank the reviewer for their valuable suggestion. To validate the generalization capabilities of the A*-Thought algorithm across diverse domains such as coding, history, computer science, and law, we conducted supplementary evaluations on the LiveCodeBench and MMLU benchmarks. It is important to note that our model was trained exclusively on mathematical tasks, making these benchmarks entirely out-domain.

The results demonstrate that the model successfully learned the A*-Thought pattern, achieving higher inference efficiency and performance on non-mathematical tasks. We will incorporate these findings into the final version of the paper.

MethodsMMLULiveCodeBenchAverageACU
Acc.Len.Acc.Len.Acc.Len.
Budget: 512
QwQ-32B37.6511.940.0512.0018.8511.973.67
+ A*-Thought56.7398.574.5509.5330.6454.056.74
Budget: 1K
QwQ-32B57.4956.900.01021.9428.7989.422.90
+ A*-Thought71.9573.3011.8986.0241.9779.665.37
Budget: 2K
QwQ-32B75.21323.263.51977.5839.41650.422.38
+ A*-Thought79.0671.4124.51734.2851.81202.854.30
Budget: 4K
QwQ-32B79.51584.0512.03586.9345.82585.491.77
+ A*-Thought80.3733.9039.03044.5159.71889.213.16

For Weakness 1

We provide a deeper theoretical analysis of A*-Thought. The rationality of "BIS + A* Search" can be explained by using importance sampling. The theoretical explanation is as follows:

Given an LRM M\mathcal{M}, it can generate an extended thinking trajectory comprising NN reasoning steps, denoted as t=t(1),t(2),,t(N)\mathbf{t} = { \mathbf{t}^{(1)}, \mathbf{t}^{(2)}, \dots, \mathbf{t}^{(N)}}, and the corresponding solution s\mathbf{s} for a given question q\mathbf{q}. This process can be represented as: (t,s):=M(q).( \mathbf{t}, \mathbf{s} ) := \mathcal{M}( \mathbf{q} ). We denote s\mathbf{s} as expectation E\mathbb{E} of M\mathcal{M} about q\mathbf{q} over the distribution t\mathbf{t}: s=EtM[q]=(qt).\mathbf{s} = \mathbb{E}_ {\mathbf{t}}^{\mathcal{M}}[\mathbf{q}] = \sum ( \mathbf{q} \mathbf{t} ). Standard LRM using uniform sampling of the thinking trajectories to approximate the expectation: EtM[q]1Ni=1Nqt(i),\mathbb{E}_ {\mathbf{t}}^{\mathcal{M}} [ \mathbf{q} ] \approx \frac{1}{N} \sum_{i=1}^{N} \mathbf{q} \mathbf{t}^{(i)}, if the final solution s\mathbf{s} is correct, we can call it an unbiased estimate. When NN is large enough, the variance of s\mathbf{s} sampled on t\mathbf{t} can be scaled by a factor of NN by the variance of q\mathbf{q} sampled uniformly on t\mathbf{t}: Vt[s]=1NVt[q].\mathbb{V}_ {\mathbf{t}} [ \mathbf{s} ] = \frac{1}{N} \mathbb{V}_ {\mathbf{t}} [ \mathbf{q} ]. However, as the number of thought steps NN keeps increasing, the model will frequently switch thought modes, leading to an increase in the variance of the thought trajectory. If such thought trajectories are sampled, it tends to reduce the convergence speed of the solution estimation and increase the chance of biased estimation.

BIS + A* Search provides an approximate method for importance sampling, which will introduce a new distribution t\mathbf{t}^{\prime}, the new expectation can be derived from EtM[q]\mathbb{E}_ {\mathbf{t}}^{\mathcal{M}} [ \mathbf{q} ]: EtM[q]=(tttq)=EtM[ttq],\mathbb{E}_ {\mathbf{t}}^{\mathcal{M}} [ \mathbf{q} ] = \sum ( \mathbf{t}^{\prime} \frac{\mathbf{t}}{\mathbf{t}^{\prime}} \mathbf{q} ) = \mathbb{E}_ {\mathbf{t}^{\prime}}^{\mathcal{M}} [ \frac{\mathbf{t}}{\mathbf{t}^{\prime}} \mathbf{q} ], the expectation of M\mathcal{M} about q\mathbf{q} over the distribution t\mathbf{t}^{\prime} can be further transformed as: EtM[ttq]1Ni=1Nqt(i)t(i).\mathbb{E}_ {\mathbf{t}^{\prime}}^{\mathcal{M}} [ \frac{\mathbf{t}}{\mathbf{t}^{\prime}} \mathbf{q} ] \approx \frac{1}{N} \sum_{i=1}^{N} \mathbf{q} \frac{\mathbf{t}^{(i)}}{\mathbf{t}^{\prime(i)}}. At this point an improvable variance about s\mathbf{s} is obtained: Vt[s]=1NVt[ttq],\mathbb{V}_ {\mathbf{t}^{\prime}} [ \mathbf{s} ] = \frac{1}{N} \mathbb{V}_ {\mathbf{t}^{\prime}} [ \frac{\mathbf{t}}{\mathbf{t}^{\prime}} \mathbf{q} ], making Vt[s]<Vt[s]\mathbb{V}_ {\mathbf{t}^{\prime}} [ \mathbf{s} ] < \mathbb{V}_ {\mathbf{t}} [ \mathbf{s} ]. Given that s\mathbf{s} is sampled on t\mathbf{t}^{\prime}, thus it can be achieved by minimizing the variance of t\mathbf{t}^{\prime}: t=argmintVt[s]=argmint1NVt[ttq]\mathbf{t}^{\prime} = \arg\min_{\mathbf{t}^{\prime}} \mathbb{V}_ {\mathbf{t}^{\prime}} [ \mathbf{s} ] = \arg\min_{\mathbf{t}^{\prime}} \frac{1}{N} \mathbb{V}_{\mathbf{t}^{\prime}} [ \frac{\mathbf{t}}{\mathbf{t}^{\prime}} \mathbf{q} ] Therefore, the ultimate goal is to find a proposal BIS distribution for the thought steps that minimizes this variance. By using the quality and information density of thought steps as heuristic information, the A* Search algorithm dynamically searches for the most important intermediate thought steps. This improves the quality of the final estimate while speeding up its convergence.

最终决定

This paper introduces A*-Thought, a framework that compresses lengthy Chain-of-Thought (CoT) outputs from large reasoning models using A* search guided by a bidirectional importance score (BIS). The method addresses efficiency challenges in verbose CoT reasoning by identifying high-value reasoning steps through attention statistics and token-level negative log-likelihood, aiming to retain performance while significantly reducing token output length.

The paper addresses a highly relevant problem of excessive reasoning length in large reasoning models, which aligns with growing interest in cost-efficient inference. The methodological innovation of integrating bidirectional importance scoring with A* search to identify essential reasoning steps is novel and well-motivated. Experimental results demonstrate substantial performance gains, with up to 2.39× improvement for QwQ-32B under constrained budgets and approximately 50% reduction in output tokens. The approach shows strong empirical performance across multiple mathematical reasoning benchmarks and demonstrates generalization capability across different large reasoning models. Thus I recommend to accept this paper.

Following the discussion phase, all major reviewer concerns have been adequately addressed. The authors successfully clarified the theoretical foundations, provided additional analysis of computational costs, and demonstrated broader applicability across different model scales. The novel combination of bidirectional importance scoring with A* search represents a valuable contribution to efficient reasoning in large language models, with strong empirical validation supporting its practical utility.