PaperHub
5.8
/10
Poster4 位审稿人
最低5最高8标准差1.3
8
5
5
5
4.0
置信度
ICLR 2024

Escape Sky-high Cost: Early-stopping Self-Consistency for Multi-step Reasoning

OpenReviewPDF
提交: 2023-09-23更新: 2024-03-14
TL;DR

We propose early-stopping self-consistency (ESC), a simple and scalable process which is capable of greatly reducing the cost of SC while maintaining performance.

摘要

关键词
Self-consistencyChain-of-ThoughtsMulti-Step ReasoningLarge Language Models

评审与讨论

审稿意见
8

To address the issue of high costs associated with self-consistency (SC), this paper introduces an approach called Early-Stop Self-Consistency (ESC). ESC incorporates an early-stop strategy into SC to reduce the overall number of samples needed. The method achieves this by dividing the large sample size used in SC into smaller sequential windows, and it stops sampling when all answers within a window are the same. Additionally, the paper presents a control scheme for ESC that dynamically selecting the size of window and maximum sampling times for different tasks and models. The effectiveness and reliability of ESC are supported by solid theoretical guarantee and extensive experiments. The empirical results demonstrate that ESC significantly reduces the average number of samples required in SC across six benchmark tasks, all while maintaining comparable performance levels.

Contributions: (1) This paper introduces an early-stop self-consistency method (ESC) to significantly reduce the computational cost of self-consistency while maintaining comparable performance. (2) This paper also puts forth a control scheme for ESC that assists in the selection of an optimal window size and maximum sampling times, considering the sampling budget and performance requirements. (3) Furthermore, this paper offers ample theoretical evidence to uphold the reliability of ESC.

优点

(1) The method is simple and effective.

(2) It is backed by a solid theoretical foundation.

(3) Extensive experiments have been conducted to confirm its effectiveness and reliability.

缺点

(1) A related paper with a similar idea, called "Let’s Sample Step by Step: Adaptive-Consistency for Efficient Reasoning with LLMs" (https://arxiv.org/pdf/2305.11860.pdf), was not referenced.

(2) In Table 1, there appear to be inaccuracies in some of the results highlighted in green. For instance, in the row labeled "Lˆ-SC (GPT4)" and the column labeled "SQA," the value "(-0.27)" should actually be "(+0.87)" because the correct difference is 0.78 (81.42 - 80.55 = 0.78). Similar issues can be found in the "SQA" column. Additionally, it's puzzling that in the "SQA" dataset, Lˆ-SC outperforms SC, even though SC has a larger sample size. This phenomenon requires further explanation.

问题

Question:

(1) In the "SQA" dataset, Lˆ-SC outperforms both SC and ESC, even though SC has a larger sample size. If there are no data errors, is there any possible reasonable explanation?

Suggestion:

(1) In Table 1, the accuracy of Lˆ-SC seems decreases slightly (less than 0.5%) in more than half situations. Therefore, it might not be accurate to claim "a large margin", as the paper does, that "We also test SC with Lˆ as the sampling size (Lˆ-SC), whose accuracy drops by a large margin.".

评论

Thank you for the positive and detailed review! We are very encouraged that you found our method to be simple, effective and robust. Below we address your comments.

Comparison with Adaptive-Consistency

Thanks for reminding and we will add citations to [1] and add more discussions on the novelty of our method:

(1) Having additional control scheme

Besides basic early-stopping self-consistency (ESC), we also propose its control scheme to dynamically choose the performance-cost balance. As we all know, along with the number of samples increases, performance will increase as well. For the realistic requirements, it is more practical to choose the affordable sample size rather than just try to maximize performance. We have derived the expectation of L^\hat{L} (sampling number) and the upper bound of E(Q)\mathbb{E}(Q) (inconsistent probability). The experiments in Section 3.3 have shown that the predictions we obtain based on Equation 10 and Equation 14 are highly reliable for balancing sampling cost and voting performance.

(2) Less sampling cost

Adaptive-consistency (AC) generates samples step by step, which means each sample requires one input. Considering the demonstrations for in-context learning (usually 8 examples) have a lot of tokens, it will cost quite a portion of the budget. By contrast, ESC generates samples in multiple sampling windows, thus samples within one window can share the same input. Table 7 in updated paper shows that ESC can get higher accuracy with less sampling cost comparing with AC. A quick preview is below:

Method# prompt# completionCostAcc
SC496.92909.00.008485.69
AC4930.3813.20.008785.66
ESC1469.21220.80.005285.67

(3) No hyperparameters

It is necessary for AC to tune the hyperparameter, specifically the confidence threshold. But for proposed ESC, the stopping criterion needs no hyperparameter due to the most conservative strategy to maintain the performance, i.e., all the answers within a window are same. Thus ESC can be conducted directly for different tasks and models, without any validation set. Besides, section 3.2 has proved the robustness regrading to window size w and max sampling size L.

(4) More solid theoretical guarantees

We provide more solid theoretical guarantees than AC lying in two parts: firstly, we rigorously derive the theoretical upper bound of the probability being inconsistent with the case where sampling size is infinite (true distribution of the model predictions according to Equation 3). While AC only considers the condition with next n samples, which is less rigorous; secondly, the upper bound of ESC is much more tight than AC. According to Equation 8, the bound value is only in the interval of 4×102\leq 4 \times 10^{-2} to 2×105\leq 2 \times 10^{-5} when window size varies from 3 to 10. By comparison,the confidence threshold of AC is chosen as 95%.

Explanation and Fixing issues in Table 1

We apologize that we have filled in the inconsistent values about SQA. That is because the candidate answer for question in SQA is either 'True' or 'False', but there will be some cases that the model generates neither of them. So there are two ways to vote the final answer: one is regarding these noisy answers as another type, e.g. '-1', to be included for voting; the other is removing this type of answers and only voting 'True' or 'False'. Our first choice was to vote based on the former criterion, but later we realized that the latter one made more sense, but we have forgotten to correct some values. We have fixed this problem in the revision.

Modification of the description about Table 1

We thank you for your suggestion! We used the description "by a large margin" because we compared the performance degradation of L-SC with the change between SC and ESC. Sorry for misleading and we have modified the depiction. Here is the reason for this phenomenon: In the paper of vanilla self-consistency [2], the large improvements come from the comparison between single greedy sample and 40 samples. The author plotted the accuracy with respect to varying numbers of samples. Within a certain interval, the performance change is not significant either.

[1] Aggarwal et al., Let's Sample Step by Step: Adaptive-Consistency for Efficient Reasoning with LLMs, EMNLP 2023.

[2] Wang et al., Self-consistency improves chain of thought reasoning in language models, ICLR 2023.

评论

Dear Authors,

I appreciate your efforts in addressing the reviews.

I consider this to be a good paper, and I maintain my score of 8 points.

审稿意见
5

The paper presents a new technique called early-stopping self-consistency (ESC) aimed at improving the computational efficiency of machine learning models, particularly in the context of complex reasoning tasks. Leveraging the essence of Chain-of-thought Reasoning and Self Consistency, the authors introduce ESC as a mechanism to strike a balance between computational cost and performance. Through extensive experiments, the paper claims significant reduction in computational overhead without a noticeable drop in performance.

优点

Originality: The introduction of ESC offers a fresh perspective in the realm of efficient machine learning algorithms. Quality: The experimental setup, including testing on six benchmarks, demonstrates the thoroughness of the research. Clarity: The paper, for the most part, is well-written and concepts are explained clearly.

缺点

Comparison with State-of-the-art: It would be helpful to see direct comparisons with current state-of-the-art methods in terms of efficiency and performance. Generalizability: The paper could discuss potential limitations or scenarios where ESC might not be the optimal solution.

问题

The Early-Stop Consistency (ESC) strategy is an optimized or "pruned" version of the Self-Consistency (SC) method, and its effect is not improved compared to SC. From this point of view, the method is lack of novelty. The primary innovation of ESC lies not in a theoretical advance but in its practical utility. It addresses real-world constraints by optimizing the balance between computational expenditure and performance fidelity.

评论

We thank your insightful comments! We understand your concerns, which are also very important to us. Below we clarify:

Comparison with state-of-the-art method

We thank you for this advice, which we believe helps strengthen our results. Per your suggestion, we conduct the experiment compared with Progressive-Hint Prompting (PHP) [1], a state-of-art method for arithmetic reasoning. In the updated paper, Table 9 from Appendix A.5, ESC (sampling size is 40) gets higher accuracy than PHP (greedy search) with less sampling cost. A quick preview is below:

Method# prompt# completionSampling CostAccuracy
CoT496.972.70.000675.83
PHP6552.0360.80.007285.08
ESC1469.21220.80.005285.67

Actually, ESC is a plug-and-play technique that can be incorporated with most of state-of-art methods, given that self-consistency (SC) is usually orthogonal to them. So we think it will be more meaningful to verify its generalization ability to SOTA methods by combining them. Also taking PHP for example in Table 10, when applying our ESC to PHP, the sampling cost drops significantly while the accuracy is basically unchanged. We believe results from the study can expand the scope of the impact of ESC.

Max sampling size10203040
PHP w. SC86.3286.6486.7687.00
PHP w. ESC86.3286.6286.7786.98
L^\hat{L}6.157.839.1510.26
L^\hat{L}-PHP w. SC86.0286.2986.3286.35

Discussion of limitations

According to Table 1, the decrease in sampling number is relatively small for Llama-2 model and MATH dataset, respectively. These two findings may reveal that the effect of ESC will be less significant under very difficult tasks and low-capability models. It makes sense since the lower accuracy will lead to fewer opportunities to reach the stop criterion, and need more samples to vote the right answer.

Novelty of ESC

Although ESC can not improve the performance compared with original SC under the same max sampling size, it can be achieved with same practical sampling cost. So from this perspective, ESC can also be seen as an method to improving effectiveness because it can get higher accuracy when enlarge the max sampling size to reach the same cost. Beside, as mentioned above, the ESC is also a plug-and-play method that can be incorporated with other competitive methods. Although ESC is more about practical significance, we provide solid theoretical guarantees to support it with rigorous upper bound, making it a key contribution in this field.

[1] Zheng et al., Progressive-hint prompting improves reasoning in large language models, arXiv.2304.09797.

审稿意见
5

The paper proposes to use an early stopping criterion, based on answer entropy, when sampling alternative answers from a LLM in a self-consistency (SC) setting. SC is a form of ensemble answering, where multiple answers are sampled and a vote decides on the final answer. With early stopping answers are sampled window-wise iteratively until the whole window contains the same answer. Experiments show that this can reduce the number of necessary calls to a LLM while maintaining similar accuracy in reasoning benchmarks.

优点

LLMs are a popular topic currently and their execution is costly, either in monetary terms or computationally. Therefore, it is a good approach to reduce the number of calls necessary, as is proposed in the paper. It is also a positive thing that existing proven techniques and statistical approaches are re-visited and used in these settings, such as early stopping or using answer entropy as a cut-off criterion. The experimental evaluation confirms the suitability of the approach over the more exhaustive standard SC technique. Experiments are extensive and consider many facets of the proposed approach.

缺点

The contribution is not particularly strong. Early stopping or using the confidence respectively the variation in multiple answers in an ensemble of answers is a well known technique. While we have (maybe, I'm not sure) not seen this in LLM sampling, it is not a particularly strong contribution in the context of an ICLR paper.

I'm also not sure we actually need the notion of the window in the method or if other statistical measurements of the confidence resp. variability could be used to determine the cut-off point. Unfortunately, this has not been discussed.

问题

No specific questions

评论

We thank your valuable comments! We understand your concerns, which are also very important to us. Below we clarify:

Contribution of our work

Firstly, the motivation of ESC is quite straightforward and significant. Reducing inference costs is a crucial topic especially for LLMs. To this end, we propose a simple and effective method to considerably reduce the cost of SC without sacrificing performance. We conduct extensive experiments to confirm that ESC can robustly save cost considering different large models, tasks (even open-ended generation tasks), decoding settings, prompts, and combination with SOTA methods (e.g., PHP, in Appendix A.6). Although ESC is simple and intuitive, we provide solid theoretical guarantees for it, which means that ESC is not a naive and heuristic method but backed with rigorous upper bound. Beyond that, we further develop a control scheme to dynamically choose the performance-cost balance according to the sampling budget and performance requirements, which is also important for realistic applications.

Discussion on the choice of introducing observation window for the design of early-stopping strategy

We want to firstly clarify that using the consistency of samples within the window as cut-off point is also a strategy based on variability.

We design the stopping strategy with the introduction of window for the following two reasons:

(1) We break the sampling process only if all of the samples in the latest window are consistent, thus avoiding any hyper-parameter. If sample one by one and stop based on the observation of the sampled samples, obviously we cannot adopt such a strict truncation condition. In this case, we need to introduce a certain statistic and its corresponding threshold (hyper-parameter), which is hard to be determined in prior.

(2) We have actually considered using the normalized entropy of the sampled samples as the statistical value for cut-off. As shown in Table below (Figure 5 in the revised version for details) , we found that this method not only has the hyper-parameter problem mentioned above, but also has no advantage over ESC in terms of performance-cost tradeoff (they both outperform SC). We believe this is because examining the cut-off point after each single sampling is too frequent, introducing greater randomness. This makes the model more likely to early-stop without sufficient sampling.

Considering the above reasons, we chose the strategy presented. We have added this discussion in the revised version (A.4), thank you for your comments.

评论

Dear authors,

thank you for taking the time and addressing the reviews, including the revision of your paper.

I acknowledge the response and improvements in the paper towards my comments.

审稿意见
5

This paper presents Early-Stopping Self-Consistency (ESC), an adaptation of the original self-consistency to reduce the sampling cost. Instead of generating all samples at once, ESC generates samples in multiple sampling windows, and stops when all samples inside the same window produce the same results. They also provide a theoretical analysis on the sampling cost and the ESC performance compared to SC. They empirically evaluate ESC on multiple reasoning benchmarks, and demonstrate that ESC achieves comparable accuracies to SC, while the number of samples notably reduces.

优点

  1. ESC is a simple yet effective adaptation of the original self-consistency to reduce the sampling cost.

  2. The ablation studies and theoretical analysis show that ESC is generally applicable to different benchmarks, and stays effective with different setups.

缺点

  1. The novelty of this work is unclear. [1] already proposed an adaptation of self-consistency to reduce the sampling cost, but this work did not cite and discuss this prior work. Without a thorough discussion and direct comparison, it is unclear whether ESC is more effective.

  2. In Table 1, when comparing ESC and L-SC, the performance difference is generally small. The reason can be that the improvement of SC saturates when the sample size increases, thus reducing the sampling size also does not drastically degrade the performance for SC. It is helpful to show this comparison for smaller sampling sizes, e.g., those in Table 2, and see if the performance improvement achieved by ESC can be more significant.

  3. There are some issues in Table 1. For example, the SQA results of L-SC are generally much higher than SC, which look problematic. Also, it is confusing to list L in the table without additional notes, as L represents the sample size, while all other rows represent the task accuracies.

[1] Aggarwal et al., Let's Sample Step by Step: Adaptive-Consistency for Efficient Reasoning with LLMs, EMNLP 2023.

问题

  1. Please clarify the novelty of this work. In particular, discuss and compare the approach to [1].

  2. Show this comparison in Table 1 for smaller sampling sizes, e.g., those in Table 2, and see if the performance improvement achieved by ESC can be more significant.

  3. Explain and fix issues in Table 1. For example, the SQA results of L-SC are generally much higher than SC, which look problematic. Also, it is confusing to list L in the table without additional notes, as L represents the sample size, while all other rows represent the task accuracies.

[1] Aggarwal et al., Let's Sample Step by Step: Adaptive-Consistency for Efficient Reasoning with LLMs, EMNLP 2023.

评论

Thanks for your detailed feedback! We will address your comments below:

Comparison with Adaptive-Consistency and the novelty of our work

Thanks for reminding and we will add citations to [1] and add more discussions on the novelty of our method:

(1) Having additional control scheme

Besides basic early-stopping self-consistency (ESC), we also propose its control scheme to dynamically choose the performance-cost balance. As we all know, along with the number of samples increases, performance will increase as well. For the realistic requirements, it is more practical to choose the affordable sample size rather than just try to maximize performance. We have derived the expectation of L^\hat{L} (sampling number) and the upper bound of E(Q)\mathbb{E}(Q) (inconsistent probability). The experiments in Section 3.3 have shown that the predictions we obtain based on Equation 10 and Equation 14 are highly reliable for balancing sampling cost and voting performance.

(2) Less sampling cost

Adaptive-consistency (AC) generates samples step by step, which means each sample requires one input. Considering the demonstrations for in-context learning (usually 8 examples) have a lot of tokens, it will cost quite a portion of the budget. By contrast, ESC generates samples in multiple sampling windows, thus samples within one window can share the same input. Table 7 in updated paper shows that ESC can get higher accuracy with less sampling cost comparing with AC. A quick preview is below:

Method# prompt# completionCostAcc
SC496.92909.00.008485.69
AC4930.3813.20.008785.66
ESC1469.21220.80.005285.67

(3) No hyperparameters

It is necessary for AC to tune the hyperparameter, specifically the confidence threshold. But for proposed ESC, the stopping criterion needs no hyperparameter due to the most conservative strategy to maintain the performance, i.e., all the answers within a window are same. Thus ESC can be conducted directly for different tasks and models, without any validation set. Besides, section 3.2 has proved the robustness regrading to window size w and max sampling size L.

(4) More solid theoretical guarantees

We provide more solid theoretical guarantees than AC lying in two parts: firstly, we rigorously derive the theoretical upper bound of the probability being inconsistent with the case where sampling size is infinite (true distribution of the model predictions according to Equation 3). While AC only considers the condition with next n samples, which is less rigorous; secondly, the upper bound of ESC is much more tight than AC. According to Equation 8, the bound value is only in the interval of 4×102\leq 4 \times 10^{-2} to 2×105\leq 2 \times 10^{-5} when window size varies from 3 to 10. By comparison,the confidence threshold of AC is chosen as 95%.

Experiments with smaller sampling size

We thank you for your insightful suggestion and we conduct experiments with smaller sampling sizes (L). From the results in updated paper, Table 8 in Appendix A.3, the absolute value of performance drop is still not significant. We believe there are two reasons for this phenomenon: (1) In the paper of vanilla self-consistency [2], the large improvements come from the comparison between single greedy sample and 40 samples. The author plotted the accuracy with respect to varying numbers of samples. Within a certain interval, the performance change is not significant either. (2) While the performance drop of Table 1 remains insignificant, the corresponding reduction of L decreases a lot compared with table 1, which proves that the interval with smaller L is more unsaturated. A quick preview is below:

MethodCSQA L=10CSQA L=20GSM8K L=10GSM8K L=20
SC77.6377.9384.1085.15
ESC77.6377.9184.1085.15
L^\hat{L}6.478.607.0910.14
L^\hat{L}-SC77.3977.5583.3184.10
评论

Fixing issues in Table 1

We apologize that we have filled in the inconsistent values about SQA. That is because the candidate answer for question in SQA is either 'True' or 'False', but there will be some cases that the model generates neither of them. So there are two ways to vote the final answer: one is regarding these noisy answers as another type, e.g. '-1', to be included for voting; the other is removing this type of answers and only voting 'True' or 'False'. Our first choice was to vote based on the former criterion, but later we realized that the latter one made more sense, but we have forgotten to correct some values. We have fixed this problem in the revision.

As for the representation of L in Table 1, thanks for your advice and we have modified Table 1 to make it more reasonable for reading.

[1] Aggarwal et al., Let's Sample Step by Step: Adaptive-Consistency for Efficient Reasoning with LLMs, EMNLP 2023.

[2] Wang et al., Self-consistency improves chain of thought reasoning in language models, ICLR 2023.

评论

Thank the authors for the response and additional experiments.

评论

We appreciate the time and effort the reviewers and AC have put into reviewing our paper. We have thus submitted our responses to the reviewers' comments and questions. We hope to receive more feedback in the discussion to further improve the paper. Please take a moment to read our responses and let us know if you have any further comments/questions. Thanks.

评论

We appreciate the reviewers' effort in reviewing our paper. Their insightful comments can make our paper more solid and comprehensive. Below is a summary of new results and updates made according to the opinions from all reviewers. Please refer to each response for details.

Add comparison with Adaptive-Consistency (AC) in Appendix A.2 and Related Work

(per request from reviewer qSHs and UrCF)

Modify description about Table 1 and add experiments with smaller sampling size in Appendix A.3

(per request from reviewer qSHs and UrCF)

Fix issues in Table 1

(per request from reviewer qSHs and UrCF)

Discuss on the choice of introducing observation window for the design of early-stopping strategy in Appendix A.4

(per request from reviewer J879)

Add comparison with state-of-the-art reasoning method in Appendix A.5 and show that ESC is actually an orthogonal method to other strong reasoning baselines in Appendix A.6

(per request from reviewer rrBP)

We do hope that we have addressed all the comments. Since there is limited time for author-reviewer discussions, please discuss with us as soon as possible if there are any further questions or comments. Thanks again for valuable time and feedback!

AC 元评审

This work proposes a cost-effective strategy for self-consistency based prompting. The proposed method works pretty well in the sense that it can reduce the cost of naive self-consistency up to ~80% while pretty much retaining accuracy. Most reviewers and I have found the proposed method valuable, however, given the simplicity of approach and existence of some related ideas, the novelty of this work is not as strong. I believe this work compensates for this through its technical merits. For instance, the authors provide theoretical upper bounds of inconsistent answers (although I am not sure if the analysis/assumptions capture what happens in practice). They also provide failry thorough analysis including robustness aspect, contollable ES, and during rebuttal, integration with other prompting methods such as PHP. Overall, this is a borderline paper but I believe it can make a decent contribution to the ICLR venue.

为何不给更高分

This paper has limited technical innovation to be considered as spotlight or oral.

为何不给更低分

I am OK if the paper is rejected. It is really borderline.

最终决定

Accept (poster)