PaperHub
6.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
4
4
4
5
3.0
置信度
创新性2.5
质量2.8
清晰度3.0
重要性2.0
NeurIPS 2025

Once Upon an Input: Reasoning via Per-Instance Program Synthesis

OpenReviewPDF
提交: 2025-05-07更新: 2025-10-29

摘要

关键词
program synthesisreasoningLLMinstancechain of thoughtneuro-symbolic

评审与讨论

审稿意见
4

The paper proposes Per-Instance Program Synthesis (PIPS), a test-time reasoning framework that decides, for each input, based on an instance-level confidence metric, whether to answer with plain chain-of-thought (CoT) reasoning or to synthesize and iteratively refine a program. Specifically, when code is chosen, the LLM/reasoner generates a symbolic abstraction of the raw input, writes a program that consumes that abstraction, runs it, and receives feedback on syntax, type, hard-coded answers and other failure patterns. Across three LLMs (Gemini-2.0-Flash, GPT-4.1-mini, o4-mini) and 30 benchmarks (23 BBEH tasks plus CLEVR, Leaf, CLUTTR and four OmniMath tasks) PIPS lifts harmonic-mean accuracy by up to 8.6% over PoT and cuts undesirable programs by 65.1% .

优缺点分析

Strengths

  1. The paper clearly articulates three core limitations of instance-level code generation: when to invoke code, absence of formal specifications, and unstructured natural language inputs. The proposed pipeline is well-aligned with these issues.
  2. The approach is evaluated across diverse reasoning domains, such as language, vision-language, relational, and mathematical reasoning tasks, with comprehensive ablations on synthesis iterations, symbolic inputs, and code-switching heuristics.
  3. The authors commit to releasing prompts and compute details. The main paper also includes a section for limitations and societal considerations, demonstrating transparency.

Weaknesses

  1. Both conceptually and empirically, authors mainly discuss the proposed approach in comparision with CoT and PoT, overlooking more recent and relevant works such as [1] (which enhances the effectiveness of solving multi-step reasoning problems by using code prompts as intermediate prompts) and [2] (which curates a library of high-level code templates (“thought-templates”) distilled from solved problems; the proposed approach retrieves & instantiates a template, then executes it).
  2. All three LLMs used (o4-mini, Gemini-2.0-Flash, GPT-4.1-mini) are proprietary. The lack of open-source LLMs limits reproducibility and transparency.
  3. Holding out 20% of every task for calibration may inflate reported accuracy and is unrealistic for truly zero-shot settings.
  4. Efficiency aspect remains under-explored. Runtime overhead of iterative synthesis (plus interpreter calls) is not reported; Figure 6 shows accuracy vs iterations but not explicit numbers (e.g., token consumption and runtime).
  5. While the paper acknowledges potential errors in converting inputs to structured symbolic form (e.g., JSON) in Section 3.3, it offers no quantitative assessment of semantic fidelity or its impact on downstream performance.
  6. In Lines 195-196, the authors report that overall, "a 0.8% absolute improvement over PoT for GPT-4.1-mini". This improvement is trivial, and raises concerns about the approach’s diminishing returns as models improve. The paper should discuss this issue more explicitly.

[1] When Do Program-of-Thought Works for Reasoning? AAAI 2024.

[2] Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models. NeurIPS 2024 Spotlight.

问题

Please refer to the weaknesses listed in Strengths And Weaknesses and address all.

局限性

yes

最终评判理由

The author's additional experiments have addressed my concerns. I am therefore inclined to raise my rating to 4.

格式问题

N/A

作者回复

We thank the reviewer for the thoughtful comments. We address the reviewer’s specific concerns in detail below.

Baselines

While the paper [1] does not present a zero-shot method for LLM reasoning, we have added a comparison with Buffer of Thoughts [2] as suggested by the reviewer. This method uses other test problems from the same task to help with determining an effective method for solving other problems of the task. Overall, we find that this is highly ineffective for BBEH problems, performing even worse than just CoT or PoT alone.

DatasetBuffer of Thoughts (gemini-2.0-flash)
Word sorting0.463
Dyck languages0.056
Object counting0.006
Object properties0.000
Boardgame qa0.275
Boolean expressions0.031
Buggy tables0.000
Spatial reasoning0.113
Multistep arithmetic0.000
Geometric shapes0.006
Temporal sequence0.000
Causal understanding0.362
Time arithmetic0.100
Web of lies0.087
Sarc triples0.194
Hyperbaton0.013
Nycc0.019
Sportqa0.125
Linguini0.006
Movie recommendation0.013
Zebra puzzles0.069
Harmonic Mean0.026

We note that we also included an agentic baseline in our original submission, the Gemini code interpreter, which we described on line 188. For another agentic baseline, we have included the smolagents implementation of a CodeAct agent, and we provide the results for all BBEH datasets using Gemini 2.0 Flash below. The complete results for all models will be included in the camera ready.

DatasetCodeAct (gemini-2.0-flash)
Word sorting0.525
Dyck languages0.044
Object counting0.000
Object properties0.006
Boardgame qa0.325
Boolean expressions0.256
Buggy tables0.025
Spatial reasoning0.287
Multistep arithmetic0.000
Geometric shapes0.369
Temporal sequence0.000
Disambiguation qa0.469
Causal understanding0.512
Time arithmetic0.256
Web of lies0.206
Sarc triples0.331
Hyperbaton0.025
Nycc0.106
Sportqa0.212
Linguini0.125
Movie recommendation0.388
Shuffled objects0.006
Zebra puzzles0.031
Harmonic Mean0.040

[1] When Do Program-of-Thought Works for Reasoning? AAAI 2024. [2] Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models. NeurIPS 2024 Spotlight.

LLMs Used

Following the reviewer’s suggestion, we have evaluated the Qwen3-235B-A22B-Instruct-2507 model as an open-weights model. We report preliminary results over all of the BBEH tasks for this model and will include the complete results in the camera ready version. We observe a 1.3% absolute improvement in harmonic mean accuracy.

DatasetPoTCoTPIPS
Word sorting0.6690.7500.750
Dyck languages0.1690.4190.456
Object counting0.5440.6560.644
Object properties0.4750.3380.331
Boardgame qa0.8440.7380.738
Boolean expressions0.3250.6000.594
Buggy tables0.4000.1690.231
Spatial reasoning0.5500.5250.525
Multistep arithmetic0.5440.5310.494
Geometric shapes0.1750.2440.244
Temporal sequence0.3380.2940.300
Causal understanding0.4500.6060.606
Time arithmetic0.7620.5690.738
Web of lies0.6130.8310.831
Sarc triples0.1810.2870.287
Hyperbaton0.2560.2870.287
Nycc0.1690.1310.131
Sportqa0.1630.2810.281
Linguini0.1440.1810.181
Movie recommendation0.7500.7250.725
Zebra puzzles0.5810.7060.706
Harmonic Mean0.3260.3660.379

20% Holdout Set for Calibration

We thank the reviewer for bringing this up. We agree that for a strict zero-shot setting, the use of a validation set is not very standard. We only use the validation set for the switch and do not use any of the validation samples in any of the prompts or for improving the downstream CoT or synthesis methods. It is possible to implement the switch without any validation samples by just using the value of the tenth question rather than training a classifier to combine all ten questions. This zero-shot switch correctly chooses between CoT and synthesis 60.3% of the time for the problems where the outcome correctness of CoT and synthesis differs. Since choosing the method which performs the best between CoT and synthesis on these problems results in only 51.2% accuracy, this result is much better than both CoT or synthesis alone, but slightly worse than the trained switch which achieves 65% accuracy. We also evaluate a leave-one-out evaluation setting where the switch used on one dataset is only trained on the 20% validation samples from other datasets. This method, called “PIPS w. LOO” shows that the switch generalizes to held out datasets rather than just held out samples.

We will include results for the setting with a switch implemented without any validation samples in our revised paper as well as the leave-one-out ablation. The PIPS w. LOO method still outperforms all baselines. For all models except GPT 4.1-mini (which already had the weakest performance), PIPS with a zero-shot switch still significantly outperforms the strongest baselines.

ModelsPIPS w. ZS SwitchPIPS w. LOOPIPS
o4-mini0.3890.3930.397
Gemini 2.0 Flash0.1990.1800.208
GPT 4.1-mini0.2590.3020.305

Cost Overhead of PIPS

We have calculated the average input and output tokens as well as average cost for each baseline method and PIPS for Gemini 2.0 Flash. We find that the iterative approaches (CodeAct and Buffer of Thoughts) increase cost over 10X compared to PoT or CoT, but PIPS is overall only 3-4X more expensive than CoT or PoT while achieving much greater accuracy.

MethodAvg. Input TokensAvg. Output TokensCost (USD)
PoT1115.961333.98$0.0006
CoT1099.771475.87$0.0007
CodeAct80137.924023.36$0.0096
Buffer of Thoughts340927.19123655.35$ 0.0835
PIPS11839.092805.78$0.0023

Quantitative Assessment of Semantic Fidelity of Input to Symbolic Mapping

Since our method first abstracts the raw input into a symbolic form, represented as JSON, there may be issues in this raw-to-symbolic mapping, as noted by the reviewer. Our preliminary investigation revealed that the vast majority of problems occur in the code itself (shown in Figure 3) and the model is often very good at the input to symbolic form translation. Now that we have addressed some of the most common problems, we think finer-grained issues should be studied in greater depth in future work, such as carrying out annotation efforts to determine the fidelity of the raw input to symbolic form mapping.

Low Improvement for GPT 4.1-mini

Our method results in only a 0.8% absolute increase in harmonic mean accuracy for GPT 4.1-mini, but this is our worst result out of many strong results across other models. In addition, we are adding an open-weights model, Qwen3-235B-A22B-Instruct-2507 (results are included above), where we see better results.

We believe PIPS has a low effectiveness with GPT 4.1-mini because the model struggles at revising generated code based on the structural feedback we provide, due to its overall weak performance. For o4-mini, a model which performs better overall in terms of the baselines, we see that PIPS improves over PoT performance by 5.7% absolute harmonic mean accuracy. Therefore, we still see that stronger models significantly benefit from using PIPS.

评论

I appreciate the authors' efforts in conducting the additional experiments and will revise my score accordingly.

I recommend that all additional results be incorporated into the paper.

评论

Hi Reviewer JESL,

Thanks again for your review. We believe we have addressed all your concerns and will include in our revision the two additional baselines (CodeAct and Buffer of Thoughts), an open-weights model, and the cost overhead information which we discussed in our rebuttal. If there are any remaining questions or concerns, we’d be grateful for your feedback.

审稿意见
4

This work introduces Per-Instance Program Synthesis ("PIPS") which aims to increase accuracy on a variety of reasoning tasks. Key elements include the detection of whether plain CoT is sufficient, or whether code should be written, with additional quality checks being applied to the code to ensure that it doesn't fall into a number of reasoning traps that the authors identified from other approaches.

优缺点分析

One key element of PIPS seems to be the learned logistic classifier over features output by asking the LLM itself about its confidence in its own capabilities at answering the question with/without PoT. While practical to implement, it doesn't seem as if this classifier achieves good accuracy in its determinations - which begs the question of whether there could be a simple probe (for a model with open weights, of course) that would perform better.

The investigation of failure modes for the code creation is of practical relevance - with the identification of 'failed structural properties' being a novel twist.

However, it seems that (from a quick look at the provided code) that the PoT implementation does not allow for regeneration/fixing of code errors, whereas PIPS is allowed re-tries. This means that the PIPS method isn't purely benefiting from the structural property guardrails, but also from the ability to fix simple typo-style errors. Choosing weak versions of the competitor baselines (as appears to be the case here, please correct me if I didn't find the correct code) isn't a good look.

问题

You state "our confidence metric allows us to correctly switch between CoT and program synthesis for 65% of cases" : As noted above, this seems as if it is a fairly weak classifier, somewhat undermining its significance. Can I confirm that the results for the methods (other than PIPS) were also conducted on the 80% of the benchmark sets that were not held out as calibration data for the logistic classifier (L139) ?

Am I correct in suggesting that the PoT implementation did not allow for code retries/rewrites when there were errors, whereas PIPS was given the opportunity for multiple self-correction loops (as illustrated in Figure 6)?

局限性

yes

最终评判理由

I'm raising, but fairly reluctantly : The method clearly works, and the experimental work is fine (and improved through the review process). OTOH, this is still fundamentally a tweak to chose between two styles of model question answering - finessing on an analysis of model errors, rather than adding a new capability into the mix

格式问题

None

作者回复

We thank the reviewer for the thoughtful comments. We address the reviewer’s specific concerns in detail below.

Confidence Metric

For our confidence-based switch between CoT and code synthesis, the 65% reported accuracy is only over the problems where exactly one of synthesis or CoT gets the correct answer, so this excludes cases where both are correct or both incorrect. Out of these problems where the outcome of CoT and synthesis differ, the best performing method out of the two gets 51.2% accuracy, so 65% accuracy is a significant improvement over both CoT and synthesis. There is still room for improvement in this switch (possibly through probe-based approaches as suggested by the reviewer), but we wanted an approach that would be easy to apply to new models, which to-date often do not provide hidden state access.

Evaluation of Baselines on 80% Non-Heldout Data

The reviewer is correct that we evaluated all methods on the 80% of the data that was not included in the heldout set used for calibration.

Iterative Baselines

The reviewer is correct that the PoT baseline did not allow for regeneration and fixing of code errors. The original version of PoT and other related methods such as FCoT [1] and PAL [2] also do not reprompt the model if there are execution errors. To address this concern, we re-run PoT with retries. Addionally, in our experiments, we included the Gemini code interpreter as a baseline which can iteratively write and debug code, as mentioned on line 188. For another such baseline, we have added the smolagents library implementation of CodeAct [3], which we find performs poorly on BBEH with gemini-2.0-flash. We will include the CodeAct results for all models in the camera ready version as well as Buffer of Thoughts (a baseline suggested by another reviewer).

DatasetCodeAct (gemini-2.0-flash)
Word sorting0.525
Dyck languages0.044
Object counting0.000
Object properties0.006
Boardgame qa0.325
Boolean expressions0.256
Buggy tables0.025
Spatial reasoning0.287
Multistep arithmetic0.000
Geometric shapes0.369
Temporal sequence0.000
Disambiguation qa0.469
Causal understanding0.512
Time arithmetic0.256
Web of lies0.206
Sarc triples0.331
Hyperbaton0.025
Nycc0.106
Sportqa0.212
Linguini0.125
Movie recommendation0.388
Shuffled objects0.006
Zebra puzzles0.031
Harmonic Mean0.040
DatasetBuffer of Thoughts (gemini-2.0-flash)
Word sorting0.463
Dyck languages0.056
Object counting0.006
Object properties0.000
Boardgame qa0.275
Boolean expressions0.031
Buggy tables0.000
Spatial reasoning0.113
Multistep arithmetic0.000
Geometric shapes0.006
Temporal sequence0.000
Causal understanding0.362
Time arithmetic0.100
Web of lies0.087
Sarc triples0.194
Hyperbaton0.013
Nycc0.019
Sportqa0.125
Linguini0.006
Movie recommendation0.013
Zebra puzzles0.069
Harmonic Mean0.026

Finally, we perform an additional experiment where we resample a potential solution using PoT whenever the generated code results in a code related error or a null result. The following shows the harmonic mean accuracy over all BBEH tasks using Gemini 2.0 Flash as the base model. We see that resampling is minimally beneficial by itself.

MethodsHarmonic Mean Acc.
PoT0.095
PoT (with retries)0.098

[1] Lyu, Qing, et al. "Faithful chain-of-thought reasoning." IJCNLP-AACL 2023. [2] Gao, Luyu, et al. "Pal: Program-aided language models." ICML 2023. [3] Wang, Xingyao, et al. "Executable code actions elicit better llm agents." ICML 2024.

评论

Thanks for running these additional experiments (even though they are reflective of what would have been a fair baseline comparison in the first place).

Having read over all the reviews/rebuttals, though, I am inclined to leave my Rating unchanged.

评论

Thank you for the follow-up. We believe we addressed your concerns regarding fair baselines and the switch performance in the rebuttal. To summarize, the additions we will include in the revision:

  • Iterative Baselines: Our original submission already included the Gemini Code Interpreter (iterative). We will add PoT+retries (which yields only a modest gain over PoT; e.g., HM 0.095→0.098 on BBEH with gemini-2.0-flash) and report CodeAct and Buffer-of-Thoughts results for all models.
  • Cost: We will additionally include the cost results in the revision (copied below) which show that PIPS achieves the highest accuracy in three times less cost than CodeAct and 35 times less cost than Buffer of Thoughts. This shows that PIPS is more cost effective than other iterative baselines. | Method | Avg. Input Tokens | Avg. Output Tokens | Cost (USD) | |-----------|-------------------|--------------------|------------| | PoT | 1115.96 | 1333.98 | 0.0006CoT1099.771475.870.0006 | | CoT | 1099.77 | 1475.87 | 0.0007 | | CodeAct | 80137.92 | 4023.36 | 0.0096BufferofThoughts340927.19123655.350.0096 | | Buffer of Thoughts | 340927.19 | 123655.35 | 0.0835 | | PIPS | 11839.09 | 2805.78 | $0.0023 |
  • Switch Significance: The reported 65% switch accuracy is computed only on instances where CoT and synthesis disagree. The best static choice on this subset is 51.2%, so our switch is a +13.8% absolute improvement.

Thank you again for pointing out the important PoT+retries baseline which we believe has significantly strengthened our experiments. If there are any other baselines which the reviewer thinks should be added, we will be happy to include them as well.

Regarding the other reviews/rebuttals, we believe we have addressed all the questions and concerns from other reviewers, but if there are any remaining concerns, we would appreciate the feedback and will try our best to respond.

审稿意见
4

This paper introduces Per-Instance Program Synthesis (PIPS), a method designed to improve the multi-step reasoning abilities of large language models (LLMs). PIPS addresses the shortcomings of existing approaches like Program of Thought (PoT), which often produce undesirable solutions.The core of PIPS is to generate and refine programs at the individual instance level, guided by structural feedback rather than explicit test cases. The method includes two key innovations: A confidence metric that dynamically decides whether to solve a problem via program synthesis or direct inference on a per-instance basis;An iterative synthesis loop that uses an evaluator to check for structural flaws like hardcoded answers, type errors, or syntax issues, providing feedback to improve the program.

优缺点分析

Strengths

  • Dynamic Adaptability: By using a confidence metric to distinguish between tasks, PIPS selects the most appropriate solution path, which helps avoid wasting computational resources on problems not suited for program synthesis.

Weaknesses

  • Questionable Generality: The method's general applicability is uncertain, as its performance improvement on certain models, such as GPT-4.1-mini, is marginal.
  • Limited Feedback Mechanism: The feedback mechanism is limited because it primarily focuses on surface-level structural issues like syntax errors and hardcoded answers. It may not be able to detect or correct deeper logical errors within a program's algorithm.
  • Computational Overhead: Compared to single-pass methods like PoT, the iterative generation and evaluation process in PIPS results in higher computational costs and increased latency.

问题

  1. The method’s upfront, binary choice between CoT and synthesis differs from tool-use agent that dynamically interleave reasoning steps with tool execution. What is the fundamental advantage of this static choice over a dynamic approach, and should the baselines have included comparisons to such agents to better situate the contributions of PIPS?

    • Yao, S., Zhao, J., Yu, D., Du, N., Shafran, I., Narasimhan, K., & Cao, Y. (2023). ReAct: Synergizing Reasoning and Acting in Language Models. International Conference on Learning Representations.

    • Paranjape, B., Korbak, F., Hall, C., Santoro, A., Lynch, C., & Hjelm, D. (2023). ART: Automatic multi-step reasoning and tool-use for large language models. arXiv preprint arXiv:2303.09014.

  2. Could the authors provide a detailed analysis of the computational overhead? Specifically, what is the average increase in end-to-end latency or total token consumption for PIPS when compared directly against the PoT baseline on the algorithmic tasks?

局限性

yes

最终评判理由

The author's supplementary experiments on multiple models (including open source models) and method overhead dispelled my concerns about the robustness of the method.

格式问题

No

作者回复

We thank the reviewer for the thoughtful comments. We address the reviewer’s specific concerns in detail below.

Method Generality

Our method results in only a 0.8% absolute increase in harmonic mean accuracy for GPT 4.1-mini, but this is our worst result out of many strong results across other models. We also emphasize that we follow the setting from the BBEH authors of using the harmonic mean accuracy over tasks which is a particularly challenging metric as it downweights easier tasks and significantly upweights the importance of the hard tasks. Even though harmonic mean accuracy improvement is low for GPT 4.1-mini, we see a 11.3% absolute improvement over PoT on Boolean Expressions and a 33.1% absolute improvement over PoT on Dyck Languages.

In addition, we are adding an open-weights model, Qwen3-235B-A22B-Instruct-2507. We report preliminary results over all of the BBEH tasks for this model and will include the complete results in the camera ready version. We observe a 1.3% absolute improvement in harmonic mean accuracy.

DatasetPoTCoTPIPS
Word sorting0.6690.7500.750
Dyck languages0.1690.4190.456
Object counting0.5440.6560.644
Object properties0.4750.3380.331
Boardgame qa0.8440.7380.738
Boolean expressions0.3250.6000.594
Buggy tables0.4000.1690.231
Spatial reasoning0.5500.5250.525
Multistep arithmetic0.5440.5310.494
Geometric shapes0.1750.2440.244
Temporal sequence0.3380.2940.300
Causal understanding0.4500.6060.606
Time arithmetic0.7620.5690.738
Web of lies0.6130.8310.831
Sarc triples0.1810.2870.287
Hyperbaton0.2560.2870.287
Nycc0.1690.1310.131
Sportqa0.1630.2810.281
Linguini0.1440.1810.181
Movie recommendation0.7500.7250.725
Zebra puzzles0.5810.7060.706
Harmonic Mean0.3260.3660.379

We believe PIPS has a low effectiveness with GPT 4.1-mini because the model struggles at revising generated code based on the structural feedback we provide, due to its overall weak performance. For o4-mini, a model which performs better overall in terms of the baselines, we see that PIPS improves over PoT performance by 5.7% absolute harmonic mean accuracy. This trend is repeated for Qwen3-235B-A22B-Instruct-2507, which is also more performant in terms of baselines.

Upfront Choice Between CoT and Synthesis

We decided to use an upfront choice between CoT and synthesis since we observed that most of the reasoning problems we were considering were solvable directly by code when the problem was “algorithmic”. For these cases, the problem was that the model could determine that the problem was solvable via program synthesis, but when prompted to generate the code, it failed to produce well-formed code. Since this was the case, even for a system that can interleave code and CoT based reasoning, the problem with code synthesis would remain.

A natural generalization of PIPS is to enable a dynamic problem-solving approach, perhaps where some subproblems are solved by synthesis and others by CoT. We believe this is an interesting future direction, but we did not find it characteristic of the BBEH problems we studied.

Agentic / Iterative Baselines

We note that the paper included an agentic baseline, the Gemini Code Interpreter, which we describe on line 188. To provide another agentic baseline, we have additionally evaluated the smolagents library implementation of a CodeAct [1] agent and we include the results below for Gemini-2.0-flash.

DatasetCodeAct (gemini-2.0-flash)
Word sorting0.525
Dyck languages0.044
Object counting0.000
Object properties0.006
Boardgame qa0.325
Boolean expressions0.256
Buggy tables0.025
Spatial reasoning0.287
Multistep arithmetic0.000
Geometric shapes0.369
Temporal sequence0.000
Disambiguation qa0.469
Causal understanding0.512
Time arithmetic0.256
Web of lies0.206
Sarc triples0.331
Hyperbaton0.025
Nycc0.106
Sportqa0.212
Linguini0.125
Movie recommendation0.388
Shuffled objects0.006
Zebra puzzles0.031
Harmonic Mean0.040

We find that CodeAct performs poorly due to several reasons. One issue is from multiple choice questions where the agent does not respond with a letter even when we explicitly give instructions for answer formatting. The other big problem can be seen for object counting and shuffled objects (two highly algorithmic tasks) where CodeAct performs very poorly due to approaching the problem by performing each low level computation with the Python interpreter and everything else with CoT rather than solving the high level problem with code. We will include CodeAct results for all models for the camera ready version.

We additionally include results for the Buffer of Thoughts [2] method over BBEH as suggested by another reviewer. This method uses other test problems from the same task to help with determining an effective method for solving other problems of the task. Overall, we find that this is highly ineffective for BBEH problems.

DatasetBuffer of Thoughts (gemini-2.0-flash)
Word sorting0.463
Dyck languages0.056
Object counting0.006
Object properties0.000
Boardgame qa0.275
Boolean expressions0.031
Buggy tables0.000
Spatial reasoning0.113
Multistep arithmetic0.000
Geometric shapes0.006
Temporal sequence0.000
Causal understanding0.362
Time arithmetic0.100
Web of lies0.087
Sarc triples0.194
Hyperbaton0.013
Nycc0.019
Sportqa0.125
Linguini0.006
Movie recommendation0.013
Zebra puzzles0.069
Harmonic Mean0.026

[1] Wang, Xingyao, et al. "Executable code actions elicit better llm agents, 2024." ICML 2024. [2] Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models. NeurIPS 2024.

Limited Feedback Mechanism

When we studied the failure modes of current models when generating code for solving reasoning problems, we found that the overwhelming majority of problems came from failing to generate code and the model resorting to textual reasoning. Since this was the most common issue prior to our work, our primary objective in this paper was to address this problem with a simple, but general framework. With this issue now addressed, we agree with the reviewer that more sophisticated forms of feedback will be necessary to address deeper problems such as logical errors now that the code is well-formed, and we are excited about future work in this direction.

Analysis of Computational Overhead

We include the average input and output tokens and the average cost for all methods below using Gemini 2.0 Flash with costs determined based on current Gemini API costs. We find that the iterative approaches (CodeAct and Buffer of Thoughts) increase cost over 10X compared to PoT or CoT, but PIPS is overall only 3-4X more expensive than CoT or PoT while achieving much greater accuracy.

MethodAvg. Input TokensAvg. Output TokensCost (USD)
PoT1115.961333.98$0.0006
CoT1099.771475.87$0.0007
CodeAct80137.924023.36$0.0096
Buffer of Thoughts340927.19123655.35$ 0.0835
PIPS11839.092805.78$0.0023
评论

The explanations regarding cost and model address my confusion, and I will raise my rating.

审稿意见
5

This paper introduces Per-Instance Program Synthesis (PIPS), a novel method that addresses the challenges of using large language models (LLMs) to generate and execute programs for complex reasoning tasks. The authors identify three key limitations of existing approaches like Program of Thought (PoT): the difficulty of deciding when to use program synthesis versus direct inference, the lack of task specifications to guide program generation, and the challenge of interfacing programs with unstructured input data. PIPS addresses these issues through an instance-level confidence metric that dynamically chooses between program synthesis and Chain of Thought reasoning, an iterative refinement process using structural feedback to improve program quality without requiring explicit test cases, and explicit symbolic extraction that converts unstructured input into structured program inputs before code generation. Experiments across 30 benchmarks including Big Bench Extra Hard tasks, visual reasoning, and mathematical problems using three frontier LLMs (Gemini-2.0-Flash, GPT-4.1-mini, and o4-mini) demonstrate that PIPS achieves up to 8.6% absolute improvement in harmonic mean accuracy over PoT and reduces undesirable program generations by 65.1%, while maintaining comparable performance to Chain of Thought on non-algorithmic tasks.

优缺点分析

Strengths

  • This paper proposes a simple yet effective approach to enable the flexible reasoning of LLMs, considering both programming and CoT, and deciding case by case
  • The paper addresses three fundamental challenges in per-instance program synthesis that have not been systematically tackled before: the open-domain nature of deciding when to use synthesis vs. direct inference, the lack of task specifications for program search, and the challenge of interfacing programs with unstructured data.
  • The paper demonstrates rigorous experimental methodology by evaluating PIPS across 30 diverse benchmarks, which was specifically designed to test advanced reasoning capabilities where current LLMs struggle. The evaluation spans three frontier LLMs (Gemini-2.0-Flash, GPT-4.1-mini, and o4-mini) and shows consistent improvements, with up to 8.6% absolute improvement in harmonic mean accuracy over PoT and a significant reduction in undesirable program generations by 65.1%.

Weaknesses

Overall, I appreciate this paper's simple yet effective approach, and I am very positive regarding its acceptance.

One weakness that bothers me a bit is that the paper misses the discussion regarding a few related works -- some of them might not be closely related to be used as baselines, but are worth discussing. For example, ViperGPT [1] implements a prominent framework for visual inference via Python execution, addresses remarkably similar problems by using LLMs to compose vision-and-language models into subroutines and generating Python code for visual reasoning tasks. PropTest [2] has also explored improving visual programming through property-based testing. I would suggest that the author should do a more comprehensive literature review over the related works, and update its related work section.

[1] Dídac Surís, Sachit Menon, Carl Vondrick. ViperGPT: Visual Inference via Python Execution for Reasoning.

[2] Jaywon Koo, Ziyan Yang, Paola Cascante-Bonilla, Baishakhi Ray, Vicente Ordonez. PropTest: Automatic Property Testing for Improved Visual Programming

问题

  • Could the authors try making a more comprehensive literature review to polish the related work section?

局限性

Yes.

最终评判理由

The authors' response mostly addresses my concern. I remain positive about this paper and I keep my score to show my support for its acceptance.

格式问题

No formatting concern.

作者回复

We thank the reviewer for their thoughtful comments and for recognizing the effectiveness and originality of our work. Following the reviewer’s comment about improving the breadth of our related work section, and the pointer to two relevant papers, we have revised the related work and believe it is much improved. We will add the following to the current related work section:

Approaches which prompt an LLM to produce code to solve a problem have been used in several domains beyond math and text-based reasoning questions. ViperGPT and followup work tackle visual question answering problems [1, 2], Voyager applies to game playing [3], and Code as Policies focusses on the application of robot control [4]. Recently, general systems such as CodeAct [5] and OpenCodeInterpreter [6] have been proposed for solving problems via code generation similar to the previously mentioned solutions for each task.

[1] Surís, Dídac, Sachit Menon, and Carl Vondrick. "Vipergpt: Visual inference via python execution for reasoning." ICCV 2023.

[2] Koo, Jaywon, et al. "PropTest: Automatic Property Testing for Improved Visual Programming." EMNLP Findings 2024.

[3] Wang, Guanzhi, et al. "Voyager: An Open-Ended Embodied Agent with Large Language Models." TMLR 2024.

[4] Liang, Jacky, et al. "Code as Policies: Language Model Programs for Embodied Control." ICRA 2023.

[5] Wang, Xingyao, et al. "Executable code actions elicit better llm agents." ICML 2024.

[6] Zheng, Tianyu, et al. "OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement." ACL Findings 2024.

We then additionally include CodeAct as a baseline and results for Gemini 2.0 Flash are included in the response to reviewer ki2H. We see that CodeAct significantly underperforms even PoT. The full results will be included in the camera ready version.

评论

Thanks for the response. The response mostly addresses my concern. I remain positive about this paper and I keep my score to show my support to its acceptance.

最终决定

This paper introduces Per-Instance Program Synthesis (PIPS), an approach that enhances the reasoning capabilities of large language models (LLMs) for complex tasks. PIPS addresses the limitations of existing methods, like Program of Thought (PoT), by dynamically deciding whether to use program synthesis or direct inference based on a confidence metric. Additionally, PIPS iteratively refines programs using structural feedback and converts unstructured inputs into structured program inputs through symbolic extraction.

The reviewers recognized the significance of PIPS in improving the performance of LLMs in multi-step reasoning, but they also raised concerns regarding the generalizability of PIPS, its computational overhead, and the limitations of its feedback mechanism. During the rebuttal phase, the authors effectively addressed these concerns through additional experiments and analysis.

Based on the reviewers’ feedback and the authors’ thorough responses, we recommend accepting this submission. And we urge the authors to address the reviewers' remaining concern on the significance of performance gain in their camera ready version.