PaperHub
6.0
/10
Poster3 位审稿人
最低6最高6标准差0.0
6
6
6
3.3
置信度
正确性3.0
贡献度2.0
表达2.7
ICLR 2025

Breaking Mental Set to Improve Reasoning through Diverse Multi-Agent Debate

OpenReviewPDF
提交: 2024-09-28更新: 2025-03-27

摘要

关键词
Multi-Agent DebateLarge Language ModelsMultimodal Large Language ModelsPromptingSelf-CorrectionReasoning

评审与讨论

审稿意见
6

The paper presents a new method, Diverse Multi-Agent Debate (DMAD), for enhancing the reasoning capabilities of large language models (LLMs) by incorporating diverse approaches to problem-solving within a debate framework. The approach builds on the Multi-Agent Debate (MAD) method by requiring each agent—represented by an LLM—to employ a unique reasoning strategy in order to address a shared problem. Traditional techniques, such as self-reflection, are limited by the model's fixed patterns of reasoning, and MAD likewise suffers when agents adopt similar approaches due to inherent limitations in reasoning diversity. By (manually) instructing each agent to use a distinct reasoning approach, DMAD seeks to enhance solution accuracy across problem-solving scenarios. In each debate round, agents independently solve a problem and then iteratively refine their solutions using insights from other agents' approaches. DMAD is tested on multiple models, including GPT-4o-mini, LLaMA-3-70B, LLaVA-1.6-13B, Gemini-1.5-Flash, and GPT-4o-2024-05-13, demonstrating overall positive, though incremental, performance gains across benchmarks.

优点

The paper contributes meaningfully to the ongoing exploration of multi-agent debate in LLMs by proposing a straight forward extension of the MAD approach that encourages diverse reasoning within the debate process. DMAD's design is simple and builds logically upon MAD, representing an incremental advance. The experiments are comprehensive and cover multiple relevant benchmarks, which gives a well-rounded view of DMAD’s performance. The inclusion of evaluations across various LLMs, both uni- and multimodal, is particularly nice, as it demonstrates the method's effectiveness across different architectures and contexts. The results show that reasoning techniques enforced by DMAD enhance performance on most tasks, confirming that diversity in reasoning approaches can be advantageous even if the cumulative improvement is moderate. Overall, the paper is clearly written and presents a straightforward extension to the multi-agent debate approaches in LLMs, supported by compelling experimental results.

缺点

While DMAD is a nice addition to the multi-agent debate line of work, its novelty is limited. The method represents an incremental improvement rather than a fundamentally new concept. The debate setup in DMAD may be perceived as a minor extension of the MAD framework, as it mainly focuses on modifying agents’ reasoning patterns. I would have liked to see additional investigations that more deeply alter the default setup of the debate, even if it was as “simple” as running an ablation study with different number of agents. Additionally, while DMAD shows an overall improvement, some gains are marginal or even negative, raising questions about the method’s broader applicability. Though the authors attempt to reason about cases where DMAD's performance lags behind other approaches, the study could benefit from a more thorough investigation of typical mistakes made during debating with DMAD.

问题

General Questions

  1. In Table 1, for the Alg. column and the LLaMA model, the value 72 achieved by DMAD is not the highest (c.f. MAD (All CoT) and CoT - SC); is this a typo or a correct results (in the latter case, please correct the text bolding)?
  2. Do you believe there would be merit in attempting to set up another LLM to propose reasoning strategies, instead of having them fixed? I understand this is a feat warranting a paper on its own, though I'm quite curious about such applications, as it would allow one to close the loop — having reasoning strategies being proposed and directly “evaluated” by the agents;
  3. Have you examined setups with more than 3 agents? Building on my previous question, I assume it's not as straight forward to find additional reasoning strategies that would be assigned to new debaters? Do you believe, given more reasoning strategies, that the performance of the method would scale up with the number of agents, or would it plateau, similar to what happens with the number of rounds?
  4. I got confused when reading Eq. 3 and Eq. 4; namely, history H\mathbb{H} is defined as a set of tuples, one per agent, each of which containing information pertaining to a particular agent (question, solving process, answer). Perhaps I am missing something obvious, but in Eq. 4, wouldn't it be sufficient to simply define hiHh_i \equiv \mathbb{H} because H\mathbb{H} by definition contains the information tuple for the agent ii, but also for all other agents? To quote the paper “hih_i represents the history of messages for MiM_i, extended with messages from other agents”, which to me looks exactly like H\mathbb{H}?
  5. Regarding the MATH benchmark, could you provide some intuition on why the performance degrades after 3 rounds (for GPT-4o-mini) compared to 2 rounds? Additionally, why does the LLaMA model's performance improve linearly with more rounds using the MAD method, but drops after the third round with DMAD? Do you have any insights into what kinds of mistakes are being made with additional rounds when using DMAD?

Minor Comments

  • Line 157: Did you mean to write hih_i instead of Ai,1A_{i,1}? The variable Ai,1A_{i,1} is not referenced in Eq. 1;
  • Line 314: Shouldn't this be "DMAD" instead of "MAD"?
  • Appendix Figures (e.g., 9, 10, 11, etc.): The figures are difficult to interpret due to the formatting characters. You may want to include pre-formatted versions for clarity and consider adding a footnote explaining that the agents actually receive raw markdown text, if necessary.
评论
  • Q5: Explanation about degraded performance and mistakes made with additional rounds when using DMAD. Why does LLaMA model's performance drop after the third round with DMAD?

This is mainly because of the typical mistakes discussed in the answer to Weakness 2. For example, in the 2nd round, Agent 1 and Agent 2 get the right answer while Agent 3 gets a wrong one. Then the judge with Self-Consistency gets the right answer. However, in the 3rd round, Agent 1 and Agent 2 mistakenly obtain the wrong conclusion, while Agent 3 correctly modifies its wrong answer. Then the judge with Self-Consistency gets the wrong answer. (Situation 1)

What we hope is, Agent 3 correctly revises its answer, and Agent 1 and Agent 2 retain their correct answers. Another ideal situation is, if two agents get wrong answers while the other one is right, all agents or at least two agents reach correct solutions in the next round. (Situation 2)

DMAD will more frequently experience that agents get different answers due to its diversity compared to MAD. If there are more Situation 2 than Situation 1, the performance will upgrade, otherwise, it degrades, which is why llama's performance drops after the 3rd round.

When to stop the debate?

So it is important to determine when the debate ends. Our method DMAD can be over in just 2 rounds and gets good performance, even surpassing MAD in 5 rounds. Besides, we can also design different criteria to determine when the debate ends, such as:

  1. Consistency-2: If there exist 2 agents getting the same answer, the debate ends. Otherwise, the debate continues unless reaching its maximum number of rounds.

  2. Consistency-3: If all the 3 agents get the same answer, the debate ends. Otherwise, the debate continues unless reaching its maximum number of rounds.

  3. Self-Determine: Set another model as a judge to determine whether the debate should be over. The judge can receive all agents' solutions in each round and save them in its history.

  4. Hybrid: In each round, if all agents get the same answer, the debate ends. Otherwise, use Self-Determine to judge whether the debate should end or continue.

We set the maximum number of the debate round to 5 with Gemini on ScienceQA. With the stop criteria, MAD and DMAD can get a relatively high accuracy with a low average round. Results show that the criteria of Consistency-3 is the best. However, DMAD with the stop criteria of Consistency-3 needs more overhead with nearly 5 debate rounds.

Stop CriteriaMethodAverage RoundAccuracyMethodAverage RoundAccuracy
Consistency-2MAD1.097785.09DMAD1.078884.93
Consistency-3MAD1.107685.14DMAD4.986185.32
Self-DetermineMAD1.109184.89DMAD1.606385.03
HybridMAD1.060584.84DMAD1.606385.03
Fixed RoundMAD184.24DMAD184.58
Fixed RoundMAD284.84DMAD285.57
Fixed RoundMAD385.34DMAD385.62
Fixed RoundMAD485.34DMAD485.52
Fixed RoundMAD585.44DMAD585.52
  • Minor Comments

Thanks for your careful review of our paper. We have revised the minor errors you mentioned in the first two items. We have also added an explanation that the agents actually receive raw markdown text.

Thanks again for your valuable feedback. We hope our responses can address your concerns. Looking forward to further communication with you.

评论
  • Q3: Examining setups with more than 3 agents, with automatically generated diverse methods and manually selected ones.

According to our answer to Q2, it may be difficult for the model to simply propose sufficiently high-quality and diverse strategies. So it is hard to guarantee the performance of DMAD with automatically generated strategies. Besides, manually selecting more high-quality strategies does not necessarily mean much higher diversity. Finding more sufficiently diverse reasoning methods is challenging. We calculate diversitydiversity for k=4k=4 strategies. The improvement is limited compared to k=3k=3. For LLaMA, diversitydiversity of k=4k=4 is even equal to k=3k=3 with CoT,SBP,PoT\\{CoT, SBP, PoT\\}.

ModelRs1{R_{s}}_{1}Rs2{R_{s}}_{2}Rs3{R_{s}}_{3}Rs4{R_{s}}_{4}diversitydiversity
GPT-4o-miniCoTL2MSBP-0.8471
GPT-4o-miniCoTL2MPoT-0.8643
GPT-4o-miniCoTSBPPoT-0.8657
GPT-4o-miniL2MSBPPoT-0.8557
GPT-4o-miniCoTL2MSBPPoT0.8743
LLaMA-3-70B-InstructCoTL2MSBP-0.6743
LLaMA-3-70B-InstructCoTL2MPoT-0.6600
LLaMA-3-70B-InstructCoTSBPPoT-0.7286
LLaMA-3-70B-InstructL2MSBPPoT-0.6386
LLaMA-3-70B-InstructCoTL2MSBPPoT0.7286

What's more, assuming we do have sufficiently diverse high-quality strategies, we think the performance may first increase, then stabilize, or even decrease at last when increasing agents. This is because, when there are too many agents, the strategies are so diverse that it leads to confusion. It may be more difficult for different agents to reach consensus, and more easily to make typical mistakes discussed in the answer to Weakness 2.

Nonetheless, we cannot rule out the possibility that this automatic setup is feasible and performs better, as we have not conducted systematic and comprehensive experiments to verify it. We greatly appreciate your insightful suggestion. It is worth leaving for future work exploration.

  • Q4: Confusion about Eq. 3 and Eq. 4.

The messages stored in history hh are in order. The message of the agent itself is appended into hh at first, and others afterwards. This is consistent with the original prompt of MAD for a fair comparison, which places the agent's own solution in front while others at the back. We will emphasize this in the revised version to help readers have a better understanding.

评论
  • Q1: "In Table 1, for the Alg. column and the LLaMA model, the value 72 achieved by DMAD is not the highest (c.f. MAD (All CoT) and CoT - SC); is this a typo or a correct results (in the latter case, please correct the text bolding)?"

Thanks for your feedback! We have corrected the text bolding.

  • Q2: "Do you believe there would be merit in attempting to set up another LLM to propose reasoning strategies, instead of having them fixed?"

This is an interesting idea. However, based on our experiments and experience, we're concerned that this approach may not lead to a better performance, especially on smaller and weaker models. The only advantage may be its automation. The reasons come from two aspects:

  1. Quality: The quality of the automatically generated prompt is not guaranteed. Instead, we select published prompting strategies that have been widely used and their effectiveness has been widely proven.

  2. Diversity: The diversity of the automatically generated prompt is not guaranteed. Instead, we can subjectively choose diverse strategies or use a designed metric to select.

Assuming we have KK candidate basic reasoning strategies Rii=1K\\{R_i\\}^K_{i=1} and want to select kk diverse ones, run each strategy NN times, and record the problems which RiR_{i} correctly solve at least once as PiP_{i} (Note that the definition of PiP_{i} here is different from PiP_{i} when introducing mental set). Note all problems on the measured dataset as PallP_{all}. We can define the diversity of the selected strategies {Rsi}i=1k\lbrace R_{s_i}\rbrace ^{k}_{i=1} as

diversity=Psii=1kP_all[0,1],si1,2,...K,sisjfor,,ij.diversity =\frac{\|\cup \\{{P_{s_i}\\}}^{k}_{i=1}\|}{\|P\_{all}\|}\in [0, 1], s_i \in \\{1, 2, ... \,K\\}, s_i \neq s_j \, for \\, \\, i \neq j.

This represents the proportion of total questions that the selected kk methods can answer correctly. The more diverse these methods are, the larger this proportion should be.

We run N=3N=3 times for each reasoning strategy in Rii=14=CoT,L2M,SBP,PoT\\{{R_i} \\}^{4}_{i=1}=\\{CoT, L2M, SBP, PoT\\} and select k=3k=3 strategies to calculate diversitydiversity, and test DMAD with different strategy groups on MATH with GPT-4o-mini. We can see diversitydiversity of {CoT, SBP, PoT} is the highest. Experiment results show that using the strategy group with larger diversitydiversity can get better results.

ModelRs1{R_{s}}_{1}Rs2{R_{s}}_{2}Rs3{R_{s}}_{3}diversitydiversity
GPT-4o-miniCoTL2MSBP0.8471
GPT-4o-miniCoTL2MPoT0.8643
GPT-4o-miniCoTSBPPoT0.8657
GPT-4o-miniL2MSBPPoT0.8557
LLaMA-3-70B-InstructCoTL2MSBP0.6743
LLaMA-3-70B-InstructCoTL2MPoT0.6600
LLaMA-3-70B-InstructCoTSBPPoT0.7286
LLaMA-3-70B-InstructL2MSBPPoT0.6386
MethodAlg.C&PGeom.Int. Alg.Num. Th.PreAlg.PreCalc.Average
DMAD (CoT, L2M, SBP)88.7±0.8278.0±1.4154.7±4.5549.0±5.1082.7±0.8285.7±0.8237.7±3.2768.0±1.83
DMAD (CoT, L2M, PoT)91.7±1.6381.3±2.9454.0±1.4154.7±0.8282.3±0.8287.7±2.1639.0±2.8370.1±0.65
DMAD (CoT, SBP, PoT)91.7±1.6381.0±1.4157.3±2.1653.7±0.8282.7±0.8286.3±0.8240.0±1.4170.4±0.95
DMAD (L2M, SBP, PoT)87.0±1.4181.3±5.7254.7±3.2751.3±1.6380.7±4.5585.0±1.4138.0±3.7468.3±2.13
评论
  • Weakness 3: Broader applicability.
  1. Statistical significance.

We run experiments with GPT-4o-mini and LLaMA-3-70B-Instruct on MATH 3 times to calculate the average accuracy and standard deviation. Statistical experiments demonstrate that DMAD outperforms other MAD settings. DMAD on LLaMA-3-70B-Instruct also gets better average accuracy than MAD (All PoT).

ModelsMethodsAlg.C&PGeom.Int. Alg.Num. Th.PreAlg.PreCalc.Avg.
GPT-4o-miniMAD (All CoT)91.3±2.1678.7±0.8255.3±2.1655.0±2.4582.7±1.6386.3±0.8239.7±0.8269.9±0.93
GPT-4o-miniMAD (All SBP)88.3±0.8277.7±1.6349.3±2.1644.0±3.7481.3±2.9483.7±0.8238.7±0.8266.1±0.35
GPT-4o-miniMAD (All PoT)91.3±1.6375.7±2.1649.0±3.7452.7±5.7280.7±0.8285.3±0.8239.7±0.8267.8±2.24
GPT-4o-miniDMAD91.7±1.6381.0±1.4157.3±2.1653.7±0.8282.7±0.8286.3±0.8240.0±1.4170.4±0.95
LLaMA-3-70B-InstructMAD (All CoT)72.7±2.9448.0±3.7431.3±0.8224.3±0.8240.7±2.9469.7±2.9431.0±1.4145.6±0.23
LLaMA-3-70B-InstructMAD (All SBP)69.3±4.3251.0±2.8329.3±4.3225.0±0.0042.0±3.7470.0±1.4127.0±1.4144.8±1.52
LLaMA-3-70B-InstructMAD (All PoT)66.7±5.8952.0±2.8331.0±2.4527.7±2.1649.0±4.9067.0±7.3532.7±1.6346.6±0.35
LLaMA-3-70B-InstructDMAD72.7±1.6349.0±4.2432.7±1.6329.3±2.9444.3±4.0872.3±2.1627.3±3.5646.8±1.43
  1. On smaller models.

We test {MAD (All CoT), MAD (All SBP), MAD (PoT), DMAD} on LLaMA-3-8B-Instruct on MATH. Results demonstrate DMAD is also effective on smaller models.

ModelsMethodsAlg.C&PGeom.Int. Alg.Num. Th.PreAlg.PreCalc.Avg.
LLaMA-3-8B-InstructMAD (All CoT)49.7±0.8219.3±1.6318.3±2.1615.3±0.8218.3±0.8246.7±1.6315.0±1.4126.1±0.42
LLaMA-3-8B-InstructMAD (All SBP)46.3±2.9420.7±4.9717.3±1.6315.3±0.8220.0±6.1644.3±5.3511.3±0.8225.3±1.94
LLaMA-3-8B-InstructMAD (All PoT)41.3±5.7220.0±2.8316.0±1.4113.7±4.5521.0±5.6639.7±4.9719.0±4.2424.4±1.62
LLaMA-3-8B-InstructDMAD47.7±0.8221.3±0.8221.7±0.8216.0±0.0021.3±5.7245.0±1.4115.0±1.4126.9±0.40
  1. On more challenging reasoning dataset of MMLU [1].

We supplement experiments on a more challenging benchmark, MMLU [1]. We test GPT-4o-mini on the subset "abstract_algebra" of MMLU and run 3 times. As this dataset consists of multi-choice questions and some options are not numbers, we replace PoT with Least-to-Most (L2M) [2]. DMAD also outperforms other MAD settings on this challenging multi-hop reasoning task.

MAD (All CoT)MAD (All SBP)MAD (All L2M)DMAD
Accuracy72.3±0.8279.0±1.4174.3±0.8279.7±1.63

[1] Measuring Massive Multitask Language Understanding. ICLR 2021.

[2] Least-to-Most Prompting Enables Complex Reasoning in Large Language Models. ICLR 2023.

评论

What is mental set?

In our paper, we introduce a new concept of mental set according to the psychological theory. Here we provide a specific definition for it. Denote MAD (All CoT), MAD (All SBP), and MAD (All PoT) as M1,M2M_{1}, M_{2} and M3M_{3} respectively. When using a kind of MAD method MiM_{i} to solve a problem, if all agents consistently get wrong answers in all debate rounds, we assume that MiM_{i} is unable to correctly solve the problem. Record all such problems for MiM_{i} as the set PiP_{i}, and get P=P1P2P3P=P_{1} \cap P_{2} \cap P_{3}. For a problem pPip \in P_{i}, if it satisfies pPp \notin P, we define that the problem pp causes mental set of MiM_{i}, and define pp as the mental set problem of MiM_{i}. It means although MiM_{i} constantly gets wrong solutions, the model can correctly solve the problem by changing to another method.

We record the mental set problems of each method on MATH with GPT-4o-mini, and the problems among them which at least one agent of other methods correctly solves. All methods have 3 agents and debate for 2 rounds. The results show that DMAD can more effectively solve other methods' mental set problems.

MAD (All CoT)MAD (All SBP)MAD (All PoT)
Number of mental set problems708767
Problems that MAD (All CoT) correctly solves045 (51.72%)46 (68.7%)
Problems that MAD (All SBP) correctly solves28 (40.0%)031 (46.3%)
Problems that MAD (All PoT) correctly solves49 (70.0%)51 (58.62%)0
Problems that DMAD correctly solves48 (68.6%)60 (69.0%)49 (73.1%)
  • Weakness 1: Limited novelty.

We would like to emphasize our novelty and differences compared with other methods leveraging diverse solutions.

  1. We introduce a new concept of mental set and provide a specific definition and systematic analysis as above, which is inspired by the psychological theory of mental set.
  2. Self-Contrast just contrasts diverse reasoning solutions to get better performance than simple IO solutions. However, they don't compare their method with each distinct reasoning strategy. So it's not sure if using diverse solving perspectives all outperforms completely using one perspective among them. Meta-Reasoning Prompting dynamically chooses the best reasoning method to solve problems, while it only achieves a balanced performance between various reasoning methods and is often inferior to a specific one in most cases. Different from them, we conduct a more comprehensive analysis and prove that leveraging diverse reasoning methods can outperform all single methods with the same setting.
  3. Our experiment shows that DMAD has the potential to break the mental set of single reasoning methods. For example, for the same problem, all agents in MAD (All IO) constantly get wrong answers in all 5 rounds. While in DMAD, an agent thinking with IO correctly solves the problem after receiving messages from agents thinking from other perspectives in the 2nd round. This also holds for MAD (All CoT), MAD (All SBP), MAD (All PoT), MAD (All CCoT) and MAD (All DDCoT). We state this in Section 4.3.4 in the initial submission.
  • Weakness 2: Investigation of typical mistakes made during debating with DMAD.

Apart from the failed case we discuss in Appendix, another typical mistake is that sometimes the agents may be affected by other agents' wrong solutions, changing its original correct answer to a wrong one. This is not the special case of DMAD, but a common issue of MAD, attracting many scholars to try to study how to alleviate this problem. However, this is not the core of our paper. We just introduce DMAD to help solve problems that MAD struggles with, improving the upper bound of its reasoning capability.

评论

Sincerely appreciate for your meticulous review of our paper! We thank for your valuable insights, which could be much helpful for improving and clarifying our paper. We would like to address your each concern in detail.

We summarize our responses as follows:

  1. We provide a specific and systematic definition for mental set, connected with the corresponding psychological phenomenon. (In Response 1/5)

  2. We emphasize our novelty and differences compared with other methods leveraging diverse solutions. (In Response 1/5)

  3. We discuss about typical mistakes of DMAD and MAD. (In Response 1/5)

  4. We conduct abundant experiments to prove the broader applicability of our method. (In Response 2/5)

    (1) Statistical significance. We re-conduct our experiments on MATH with GPT-4o-mini and LLaMA-3-70B-Instruct 3 times to get the average accuracy and standard deviation. Our method DMAD consistently outperforms other MAD methods.

    (2) We test on a smaller model, LLaMA-3-8B-Instruct. DMAD also achieves the best performance.

    (3) We conduct experiments on the subset of a more challenging benchmark, MMLU [1], where DMAD also outperforms other methods.

  5. We elaborate our opinions on automatically generating prompting strategies (In Response 3/5) and setting more agents (In Response 4/5).

  6. We answer the questions about equations. (In Response 4/5)

  7. We explain the degraded performance of LLaMA in the 3rd round. (In Response 5/5)

  8. We design 4 different stop criteria to determine when the debate ends. (In Response 5/5)

We also respond to other minor questions, and have added these supplementary experiments in the revised paper. We hope our responses can address your concerns. Looking forward to further communication with you.

[1] Measuring Massive Multitask Language Understanding. ICLR 2021.

评论

I thank the authors for their comprehensive response. After consideration, I decided to keep my initial assessment — the work, while incremental, would in my opinion represent an interesting and valuable addition to the community.

评论

Thanks for your reply! We deeply appreciate your insightful feedback and will continue to make progress.

审稿意见
6

This paper proposes Diverse Multi-Agent Debate (DMAD), an improved version of Multi-Agent Debate (MAD) that leverages agents using different reasoning methods to break "mental set" and enhance reasoning performance. In DMAD, each agent adopts a distinct prompting strategy (e.g. chain-of-thought, step-back prompting, program-of-thoughts) to generate solutions. The agents then review each other's solutions, extract insights from the different reasoning approaches, and iteratively refine their own responses. Experiments on both language models and multimodal language models across various benchmarks demonstrate that DMAD outperforms MAD with homogeneous reasoning as well as other baselines like self-reflection methods. Key results include DMAD in 2 rounds surpassing MAD in 5 rounds, and improvements of 7-8% on metrics compared to 5% for standard MAD. The paper provides a systematic analysis of DMAD, showing performance improves with more debate rounds and agents using a greater diversity of reasoning methods.

优点

  1. The key idea of using diversity in reasoning methods to break mental set and improve multi-agent debate is novel and creative. This can leverage different type of strneghts and activation patternsa of LLMs at inference time to produce diverse outputs. This leverages the ideas from cognitive science - the idea of mental set is interestijng.
  2. The approach is evaluated across a variety of models (multiple open and closed-source language and multimodal models), datasets/tasks (4 language and multimodal benchmarks), and against strong baselines (standard MAD, self-reflection methods, etc). The authors demonstrate that DMAD approach works well.
  3. The ablation studies on number of agents, debate rounds, importance of thought process, etc provide ways to undersstand the robustness of the papers results.

缺点

While the experiments cover a good range of models and tasks, it would have been nice to see results on more challenging reasoning datasets like MMLU to further validate the approach.

While the paper shows that using agents with diverse reasoning strategies is helpful, its not clear on what kind of diversity is most effective. The authors use three pre-defined reasoning methods (chain-of-thought, step-back, program-of-thoughts), but do not explore other possibilities or study the properties of these methods that make them complementary. A more principled analysis of reasoning diversity could be helpful.

The paper shows the benefits of DMAD with a fixed number of agents (3) and debate rounds (2-5). However, there is no systematic analysis of how the performance and computational cost scale with these parameters. In real-world applications, it may be necessary to use a much larger number of agents and rounds to tackle complex reasoning tasks. Understanding the scalability properties of DMAD is important. Also, for the multimodal experiments, it's unclear if the visual information is playing an important role. Adding experiments on more vision-centric tasks/datasets could strengthen this aspect. Its also not well discussed why self-reflection methods struggle on multimodal tasks compared to language-only tasks. Is this an inherent limitation or due to the specific self-reflection approaches used?

The paper is a good start but stops short on addressing these points.

问题

Some questions came up during the analysis:

How does DMAD perform on tasks that require multi-hop reasoning over longer contexts? Do the gains over baselines hold there as well? Expanding on the above, it would be great to see MMLU results, even if just on a subset of subjects most relevant to reasoning. Can the authors comment on the computational overhead of DMAD compared to standard MAD, both at inference time and during the debate rounds? Since multimodal models are more computationally intensive in general, were the number of trials/random seeds adjusted for those experiments? Clarifying those details would be helpful. Figure 1 is quite interesting - if space permits, visualizing even more of the ablations (number of rounds, removing step output, etc) in this format could provide further insight.

评论
  • Adding experiments on more vision-centric tasks/datasets.

In our submitted paper, we conduct experiments on both LLMs and MLLMs to further verify the effectiveness of DMAD on different modalities. We test 3 MLLMs on two vision reasoning benchmarks including multiple aspects such as object recognition, optical character recognition, knowledge answering, math vision-question answering, language generation, and spatial awareness, which requires MLLMs to solve vision-centric problems according to the given image.

We believe our evaluated datasets and tasks provide evidence to support the conclusions drawn. However, we greatly value the reviewer’s insights and would be happy to explore additional tasks or datasets that require different types of reasoning or demonstrate other properties, if the reviewer has specific recommendations.

  • Why self-reflection methods struggle on multimodal tasks compared to language-only tasks?

We discuss it in Appendix C.3. This may be due to the underconfidence of MLLMs, which tend to believe that their initial answer is incorrect and modify it, even if most of their answers are right. Here we compared Self-Refine results of MLLMs on ScienceQA and LLMs on MATH. We can find MLLMs are more likely to change their right answers to wrong ones.

TypeModelMaintainRight -> WrongWrong -> RightWrong -> WrongAccuracy Variation
MLLMLLaVA36.2438.7714.2810.71-24.49
MLLMGemini36.4948.448.286.79-40.16
MLLMGPT-4o66.0025.004.005.00-21.00
LLMLLaMA38.8614.438.7138.00-5.71
LLMGPT-4o-mini79.144.294.7111.86+0.43

(Maintain: The answer remains unchanged. Right -> Wrong: A right answer is changed to wrong. Wrong -> Right: A wrong answer is changed to right. Wrong -> Wrong: A wrong answer is changed but remains incorrect.)

However, this may be affected by specific prompts. The evaluation prompt in Self-Refine, "Review your previous answer and find problems with your answer.", and the refinement prompt, "Based on the problems you found, improve your answer.", may lead models to nitpick on correct answers and find problems of right solutions. Therefore, we change the evaluation prompt to "Review your previous answer and determine whether your previous answer is right or wrong." to get feedback, and use "Based on your judgment, improve your answer. If your previous answer is judged as wrong, modify it to be correct. Otherwise, keep your previous answer." to refine. On LLaVA and Gemini, this prompt setting performs better than Self-Refine, while still getting worse results than before modification.

TypeModelMaintainRight -> WrongWrong -> RightWrong -> WrongAccuracy Variation
MLLMLLaVA66.8317.959.026.20-8.93
MLLMGemini67.0120.387.275.34-13.14
LLMLLaMA30.7120.009.1440.14-10.86
LLMGPT-4o-mini86.571.574.007.86+2.43
  • Were the number of trials/random seeds adjusted for experiments on MLLMs?

No. On the the same benchmark and model, we strictly test the performance of different methods using completely identical settings, only the prompts are different. Although MLLMs are computationally intensive, we test LLaVA, Gemini, and GPT-4o on the whole MM-Vet benchmark. We also test LLaVA and Gemini on all ScienceQA samples, and test GPT-4o with 100 random samples with the sample seed 0 for reproduction. We clarify these in Section 4.2.2.

  • Visualizing more of the ablations in the format of Figure 1.

Thanks for your advice. We have added such a figure in Appendix.

Thanks again for your valuable feedback. We hope our responses can address your concerns. Looking forward to further communication with you.

评论
  • Computational overhead including tokens and inference time of DMAD and MAD during the debate rounds.

Thanks for your advice. This is important for real-world applications. Based on our record of experiment results on Gemini on ScienceQA, we use a word as a token to calculate the tokens overhead of traditional MAD, i.e., MAD (All IO), and DMAD during the debate rounds.

We can see that (DMAD tokens / MAD tokens) becomes larger as the debate round increases. This is because the selected diverse strategies generate more tokens. The diverse solutions are added to each agent's debate history in each round. The accumulated history of diverse solutions will cause more and more overhead. Nonetheless, DMAD in 2 rounds gets better performance and needs lower overhead and calls than MAD when MAD achieves its best performance in 5 rounds.

RoundDMAD AccuracyMAD AccuracyDMAD tokensMAD tokensDMAD tokens / MAD tokens
184.5884.241,169,209721,8001.61985
285.5784.844,114,7802,059,7911.99767
385.6285.347,753,6473,585,6692.16240
485.5285.3411,517,7265,215,0682.20855
585.5285.4415,364,8276,905,9852.22486

When to stop the debate?

Besides, it is important to determine when the debate ends to reduce additional overhead. Although we run all experiments in the fixed round for a fair comparison, which is consistent with the settings of the paper of MAD, we can also design different criteria to determine when the debate ends, such as:

  1. Consistency-2: If there exists 2 agents getting the same answer, the debate ends. Otherwise, the debate continues unless reaching its maximum number of rounds.

  2. Consistency-3: If all the 3 agents get the same answer, the debate ends. Otherwise, the debate continues unless reaching its maximum number of rounds.

  3. Self-Determine: Set another model as a judge to determine whether the debate should be over. The judge can receive all agents' solutions in each round and save them in its history.

  4. Hybrid: In each round, if all agents get the same answer, the debate ends. Otherwise, use Self-Determine to judge the debate should end or continue.

We set the maximum number of the debate round to 5. With the stop criteria, MAD and DMAD can get a relatively high accuracy with low average round. Results show that the criteria of Consistency-3 is the best. However, DMAD with the stop criteria of Consistency-3 needs more overhead with nearly 5 debate rounds.

Stop CriteriaMethodAverage RoundAccuracyMethodAverage RoundAccuracy
Consistency-2MAD1.097785.09DMAD1.078884.93
Consistency-3MAD1.107685.14DMAD4.986185.32
Self-DetermineMAD1.109184.89DMAD1.606385.03
HybridMAD1.060584.84DMAD1.606385.03
Fixed RoundMAD184.24DMAD184.58
Fixed RoundMAD284.84DMAD285.57
Fixed RoundMAD385.34DMAD385.62
Fixed RoundMAD485.34DMAD485.52
Fixed RoundMAD585.44DMAD585.52

The inference time of DMAD and MAD has the same trend as tokens. We re-conduct experiment on MATH with gpt-4o-mini and record the tokens overhead and inference time of DMAD and MAD. We run 3 times to calculate the average accuracy and standard deviation. Our DMAD approach effectively strikes a balance among different approaches that require varying amount of inference computation while achieving optimal performance. We hope the reported results can be helpful.

MethodTokensCost ($)Inference Time (s)Accuracy
MAD (All CoT)4,445,0661.507734,37669.9±0.93
MAD (All SBP)11,076,7923.541976,42566.1±0.35
MAD (All PoT)3,113,7161.021525,18967.8±2.24
DMAD6,331,3162.044945,51470.4±0.95
评论
  • What is mental set?

In our paper, we introduce a new concept of mental set according to the psychological theory. Here we supplement a specific definition for it. Denote MAD (All CoT), MAD (All SBP), and MAD (All PoT) as M1,M2M_{1}, M_{2} and M3M_{3} respectively. When using a kind of MAD method MiM_{i} to solve a problem, if all agents consistently get wrong answers in all debate rounds, we assume that MiM_{i} is unable to solve the problem correctly. Record all such problems for MiM_{i} as the set PiP_{i}, and get P=P1P2P3P=P_{1} \cap P_{2} \cap P_{3}. For a problem pPip \in P_{i}, if it satisfies pPp \notin P, we define that the problem pp causes mental set of MiM_{i}, and define pp as the mental set problem of MiM_{i}. It means although MiM_{i} constantly gets wrong solutions to pp, the model can correctly solve the problem by changing to another method.

We record the mental set problems of each method on MATH with GPT-4o-mini, and the problems among them which at least one agent of other methods correctly solves. All methods have 3 agents and debate for 2 rounds. The results show that DMAD can more effectively solve other methods' mental set problems.

MAD (All CoT)MAD (All SBP)MAD (All PoT)
Number of mental set problems708767
Problems that MAD (All CoT) correctly solves045 (51.72%)46 (68.7%)
Problems that MAD (All SBP) correctly solves28 (40.0%)031 (46.3%)
Problems that MAD (All PoT) correctly solves49 (70.0%)51 (58.62%)0
Problems that DMAD correctly solves48 (68.6%)60 (69.0%)49 (73.1%)
  • What kind of diversity is most effective?

We initially chose divergent prompting strategies by intuition. CoT solves the problem step by step, but may struggle with problems involving complex theorems. SBP first extracts relevant principles and then solves the problem according to them, but may weaken the ability to solve problems step by step. PoT uses Python programs, but this is mostly suitable for problems involving numbers. Therefore, these strategies are diverse and can complement each other.

To offer a more quantitative analysis, we design a more objectively quantitative metric. Assuming we have KK candidate basic reasoning strategies Rii=1K\\{R_i\\}^K_{i=1} and want to select kk diverse ones, run each strategy NN times, and record the problems which RiR_{i} correctly solve at least once as PiP_{i} (Note that the definition of PiP_{i} here is different from PiP_{i} when introducing mental set). Note all problems on the measured dataset as PallP_{all}. We can define the diversity of the selected strategies {Rsi}i=1k\lbrace R_{s_i}\rbrace ^{k}_{i=1} as

diversity=Psii=1kP_all[0,1],si1,2,...K,sisjfor,,ij.diversity =\frac{\|\cup \\{{P_{s_i}\\}}^{k}_{i=1}\|}{\|P\_{all}\|}\in [0, 1], s_i \in \\{1, 2, ... \,K\\}, s_i \neq s_j \, for \\, \\, i \neq j.

This represents the proportion of total questions that the selected kk methods can answer correctly. The more diverse these methods are, the larger this proportion should be.

We run N=3N=3 times for each reasoning strategy in Rii=14=CoT,L2M,SBP,PoT\\{{R_i} \\}^{4}_{i=1}=\\{CoT, L2M, SBP, PoT\\} and select k=3k=3 strategies to calculate diversitydiversity, and test DMAD with different strategy groups on MATH with GPT-4o-mini. We can see that using the method group with larger diversitydiversity can get better results.

ModelRs1{R_{s}}_{1}Rs2{R_{s}}_{2}Rs3{R_{s}}_{3}diversitydiversity
GPT-4o-miniCoTL2MSBP0.8471
GPT-4o-miniCoTL2MPoT0.8643
GPT-4o-miniCoTSBPPoT0.8657
GPT-4o-miniL2MSBPPoT0.8557
LLaMA-3-70B-InstructCoTL2MSBP0.6743
LLaMA-3-70B-InstructCoTL2MPoT0.6600
LLaMA-3-70B-InstructCoTSBPPoT0.7286
LLaMA-3-70B-InstructL2MSBPPoT0.6386
MethodAlg.C&PGeom.Int. Alg.Num. Th.PreAlg.PreCalc.Average
DMAD (CoT, L2M, SBP)88.7±0.8278.0±1.4154.7±4.5549.0±5.1082.7±0.8285.7±0.8237.7±3.2768.0±1.83
DMAD (CoT, L2M, PoT)91.7±1.6381.3±2.9454.0±1.4154.7±0.8282.3±0.8287.7±2.1639.0±2.8370.1±0.65
DMAD (CoT, SBP, PoT)91.7±1.6381.0±1.4157.3±2.1653.7±0.8282.7±0.8286.3±0.8240.0±1.4170.4±0.95
DMAD (L2M, SBP, PoT)87.0±1.4181.3±5.7254.7±3.2751.3±1.6380.7±4.5585.0±1.4138.0±3.7468.3±2.13
评论

We thank the reviewer for considering our idea of mental set is interesting and commending our sufficiency of experiments. We also appreciate the valuable and insightful feedback provided, which could be much helpful to better prove the effectiveness of our method. We would like to address each of your concerns.

We summarize our responses as follows:

  1. We conduct experiments on the subset of MMLU, where our method DMAD also outperforms other methods. (In Summary Response)
  2. We provide a specific and systematic definition for mental set, connected with the corresponding psychological phenomenon. (In Respone 1/3)
  3. We design an objectively quantitative metric to measure the diversity of selected basic prompting strategies, and conduct several experiments to prove the effectiveness of this metric. (In Response 1/3)
  4. We report the computational overhead including tokens and inference time of DMAD and MAD during the debate rounds. DMAD achieves optimal performance with balanced overhead compared with other MAD methods. (In Response 2/3)
  5. We design 4 different stop criteria to determine when the debate ends to reduce excessive tokens. (In Response 2/3)
  6. We test different MLLMs and LLMs using 2 different self-reflection prompts, proving and analyzing the poor performance of self-reflection on MLLMs. (In Response 3/3)

We also respond to other minor questions and have added these supplements in the revised paper. We hope our responses can address your concerns. Looking forward to further communication with you.

  • Experiments on more challenging multi-hop reasoning tasks such as MMLU.

Thanks for your suggestion. We use GPT-4o-mini to conduct experiments on the subset "abstract_algebra" of MMLU and run 3 times to get the average accuracy and standard deviation for each method. As this dataset consist of multi-choice questions and some options are not numbers, we replace PoT with Least-to-Most (L2M) [1]. DMAD also outperforms other MAD settings on this challenging multi-hop reasoning task.

MAD (All CoT)MAD (All SBP)MAD (All L2M)DMAD (Ours)
Accuracy72.3±0.8279.0±1.4174.3±0.8279.7±1.63

[1] Least-to-Most Prompting Enables Complex Reasoning in Large Language Models. ICLR 2023.

审稿意见
6

The pair introduces Diverse Multi-Agent Debate (DMAD), a new method that encourages AI agents to employ different reasoning approaches when debating to solve problems. Unlike traditional MAD, in which agents use the same reasoning method, DMAD assigns distinct reasoning strategies to each agent. DMAD consistently outperformed other methods across multiple benchmarks, demonstrating that diverse reasoning approaches can help break mental sets - the tendency to approach problems in fixed ways based on past experience. This work provides a promising direction for enhancing AI reasoning capabilities without requiring model retraining.

优点

1: The solution is simple since it doesn't require model retraining or additional data collection, making it immediately applicable. 2: The research provides extensive experimental validation across multiple dimensions and shows consistent improvements.

缺点

1:Resource Efficiency Analysis

The paper lacks a critical discussion on computational efficiency and practical costs. A detailed analysis comparing token usage, API costs, and computational overhead between DMAD and simpler reasoning strategies would better demonstrate its practical value. This is particularly important for real-world applications where resource constraints must be balanced against performance gains.

2: Weak Theoretical Foundation and Empirical Evidence

While the paper draws motivation from psychology's "mental set" phenomenon, the connection appears tenuous. Traditional MAD with a single reasoning method (e.g., CoT) achieves comparable performance to DMAD (as shown in Table 1), suggesting that method diversity might not be the key driver of improvement (Or statistical significance should be provided). This weakens the paper's central thesis that diverse reasoning methods are necessary to break mental set, as even fixed methods can yield different problem-solving approaches through debate. This raises questions about whether the gains are truly from method diversity or simply from the debate process itself (which also encourages some sort of "divergent thinking").

3: Model Capability Scaling

The paper's focus on large, capable models leaves an important question unexplored: could smaller models using DMAD potentially outperform larger models using simpler reasoning methods? This comparative analysis would better demonstrate the method's value proposition, especially for scenarios where using large models isn't feasible. Understanding how DMAD's effectiveness scales with model size could provide valuable insights for practical applications.

4:Limited Implementation Guidelines

Despite DMAD's potential impact on engineering workflows, the paper lacks concrete guidance on implementation decisions. Critical questions remain unanswered; See Questions blow for more.

5: Incremental Rather Than Novel Contribution:

While the paper effectively implements diverse reasoning in a MAD framework, the core concept builds on existing ideas. Similar approaches of leveraging diverse perspectives for improved reasoning have been explored in recent work (Zhang et al., 2024). Although DMAD shows impressive results, its primary contribution appears to be an incremental improvement (applying diversity to MAD) rather than introducing a fundamentally new concept to the field.

Reference: Zhang, W., Shen, Y., Wu, L., Peng, Q., Wang, J., Zhuang, Y., & Lu, W. (2024). Self-Contrast: Better Reflection Through Inconsistent Solving Perspectives. 2024 ACL.

问题

1: Is there a reason you chose the 3 reasoning strategies (CoT, SBP, PoT)? How about others like least-to-most prompting (Zhou et al., 2023)?

2: Could the system dynamically adjust its reasoning methods during the debate process? For instance, could agents switch reasoning strategies based on the progress of the debate or the nature of disagreements that emerge?

3: Why do you focus on mathematical reasoning and visual understanding? How about other tasks?

4: What criteria determine the ideal number of debate rounds? Examining Figure 4 (Performance with increased rounds), there's an interesting pattern where DMAD sometimes achieves better performance in 2 rounds than MAD does in 5 rounds, however would a smarter stopping criteria can help mitigate the problem?

Reference:

Zhou, Denny, Nathanael Schärli, Le Hou, Jason Wei, Nathan Scales, Xuezhi Wang, Dale Schuurmans, Claire Cui, Ed Chi, Yun-Hsuan Sung, Daphne Ippolito, and Claire Cassidy. 2023. "Least-to-most Prompting Enables Complex Reasoning in Large Language Models." In International Conference on Learning Representations (ICLR).

评论
  • Question3: Why focusing on mathematical reasoning and visual understanding?

We conduct abundant experiments in multiple aspects, not limited to mathematical reasoning and vision understanding. In our paper, we conduct experiments on mathematics, biology, chemistry, and physics reasoning. We also test on Scientific Vision-Question Answering (VQA) including natural, social, and language science, and MM-Vet involving various visual capabilities such as object recognition, optical character recognition, knowledge answering, math VQA, language generation, and spatial awareness. These benchmarks have been widely used by researchers.

What's more, we supplement experiments on a more challenging benchmark, MMLU [1]. We use GPT-4o-mini to conduct experiments on the subset "abstract_algebra" of MMLU and run 3 times to get the average accuracy and standard deviation for each method. As this dataset consists of multi-choice questions and some options are not numbers, we replace PoT with L2M. DMAD also outperforms other MAD settings on this challenging multi-hop reasoning task.

MAD (All CoT)MAD (All SBP)MAD (All L2M)DMAD
Accuracy72.3±0.8279.0±1.4174.3±0.8279.7±1.63
  • Question 4: What criteria determine the ideal number of debate rounds? Is there a smarter stopping criteria?

To fairly compare with MAD, we adopt all the same settings in their paper, which executes MAD in the fixed round. Without considering this, we can design different criteria to determine when the debate ends, such as:

  1. Consistency-2: If there exist 2 agents getting the same answer, the debate ends. Otherwise, the debate continues unless reaching its maximum number of rounds.

  2. Consistency-3: If all the 3 agents get the same answer, the debate ends. Otherwise, the debate continues unless reaching its maximum number of rounds.

  3. Self-Determine: Set another model as a judge to determine whether the debate should be over. The judge can receive all agents' solutions in each round and save them in its history.

  4. Hybrid: In each round, if all agents get the same answer, the debate ends. Otherwise, use Self-Determine to judge whether the debate should end or continue.

We set the maximum number of the debate round to 5 with Gemini on ScienceQA. With the stop criteria, MAD and DMAD can get a relatively high accuracy with a low average round. Results show that the criteria of Consistency-3 is the best. However, DMAD with the stop criteria of Consistency-3 needs more overhead with nearly 5 debate rounds.

Stop CriteriaMethodAverage RoundAccuracyMethodAverage RoundAccuracy
Consistency-2MAD1.097785.09DMAD1.078884.93
Consistency-3MAD1.107685.14DMAD4.986185.32
Self-DetermineMAD1.109184.89DMAD1.606385.03
HybridMAD1.060584.84DMAD1.606385.03
Fixed RoundMAD184.24DMAD184.58
Fixed RoundMAD284.84DMAD285.57
Fixed RoundMAD385.34DMAD385.62
Fixed RoundMAD485.34DMAD485.52
Fixed RoundMAD585.44DMAD585.52

Thanks again for your valuable feedback. We hope our responses can address your concerns. Looking forward to further communication with you.

评论
  • Weakness 2-1: Weak theoretical foundation and empirical evidence.

Our idea is inspired by the psychological theory of mental set, which refers to the cognitive tendency to approach problems in a particular way based on past experiences, learned behaviors, or established habits. This may hinder diverse thinking and make it difficult to solve a problem in the mental set way. However, if thinking in a different way, it may surprisingly find it is easy to address the problem. We observe that LLMs have an analogous phenomenon. MAD with a fixed prompting strategy may always get wrong answers to a problem, while changing to another prompting strategy can correctly solve the problem.

Here we provide a specific definition for it. Denote MAD (All CoT), MAD (All SBP), and MAD (All PoT) as M1,M2M_{1}, M_{2} and M3M_{3} respectively. When using a kind of MAD method MiM_{i} to solve a problem, if all agents consistently get wrong answers in all debate rounds, we assume that MiM_{i} is unable to correctly solve the problem. Record all such problems for MiM_{i} as the set PiP_{i}, and get P=P1P2P3P=P_{1} \cap P_{2} \cap P_{3}. For a problem pPip \in P_{i}, if it satisfies pPp \notin P, we define that the problem pp causes mental set of MiM_{i}, and define pp as the mental set problem of MiM_{i}. It means although MiM_{i} constantly gets wrong solutions to pp, the model can correctly solve the problem by changing to another strategy.

We record the mental set problems of each method on MATH with GPT-4o-mini, and the problems among them which at least one agent of other methods correctly solves. All methods have 3 agents and debate for 2 rounds. The results prove the connection between LLM reasoning and the mental set phenomenon in psychology, and DMAD can more effectively solve other methods' mental set problems. In the initial submission, we state that DMAD can solve mental set problems of other MAD methods in Section 4.3.4 and list many examples in Appendix D.

MAD (All CoT)MAD (All SBP)MAD (All PoT)
Number of mental set problems708767
Problems that MAD (All CoT) correctly solves045 (51.72%)46 (68.7%)
Problems that MAD (All SBP) correctly solves28 (40.0%)031 (46.3%)
Problems that MAD (All PoT) correctly solves49 (70.0%)51 (58.62%)0
Problems that DMAD correctly solves48 (68.6%)60 (69.0%)49 (73.1%)
  • Weakness 2-2: Statistical significance.

We run experiments with GPT-4o-mini and LLaMA-3-70B-Instruct on MATH 3 times to calculate the average accuracy and standard deviation. Statistical experiments demonstrate that DMAD outperforms other MAD settings. DMAD on LLaMA-3-70B-Instruct also gets better average accuracy than MAD (All PoT). This implies the gains come from method diversity rather than the debate process itself to some extent.

ModelsMethodsAlg.C&PGeom.Int. Alg.Num. Th.PreAlg.PreCalc.Avg.
GPT-4o-miniMAD (All CoT)91.3±2.1678.7±0.8255.3±2.1655.0±2.4582.7±1.6386.3±0.8239.7±0.8269.9±0.93
GPT-4o-miniMAD (All SBP)88.3±0.8277.7±1.6349.3±2.1644.0±3.7481.3±2.9483.7±0.8238.7±0.8266.1±0.35
GPT-4o-miniMAD (All PoT)91.3±1.6375.7±2.1649.0±3.7452.7±5.7280.7±0.8285.3±0.8239.7±0.8267.8±2.24
GPT-4o-miniDMAD (Ours)91.7±1.6381.0±1.4157.3±2.1653.7±0.8282.7±0.8286.3±0.8240.0±1.4170.4±0.95
LLaMA-3-70B-InstructMAD (All CoT)72.7±2.9448.0±3.7431.3±0.8224.3±0.8240.7±2.9469.7±2.9431.0±1.4145.6±0.23
LLaMA-3-70B-InstructMAD (All SBP)69.3±4.3251.0±2.8329.3±4.3225.0±0.0042.0±3.7470.0±1.4127.0±1.4144.8±1.52
LLaMA-3-70B-InstructMAD (All PoT)66.7±5.8952.0±2.8331.0±2.4527.7±2.1649.0±4.9067.0±7.3532.7±1.6346.6±0.35
LLaMA-3-70B-InstructDMAD (Ours)72.7±1.6349.0±4.2432.7±1.6329.3±2.9444.3±4.0872.3±2.1627.3±3.5646.8±1.43
评论

We appreciate your comprehensive review and valuable feedback of our paper. We would like to address each of your questions.

We summarize our responses as follows:

  1. We elaborate the connection between our idea and the psychological theory of mental set, and provide a specific and systematic definition for it, proven by our experiment results. (In Response 1/4)
  2. We re-conduct our experiments on MATH with GPT-4o-mini and LLaMA-3-70B-Instruct 3 times to get the average accuracy and standard deviation. Our method DMAD consistently outperforms other MAD methods. (In Response 1/4)
  3. We report the tokens overhead and cost of all methods on MATH. DMAD achieves optimal performance with balanced overhead compared with other MAD methods. (In Response 2/4)
  4. We explain why choosing CoT, SBP, and PoT, and also test other groups including Least-to-Most prompting. We design an objectively quantitative metric to measure the diversity of selected basic prompting strategies, and conduct several experiments to prove the effectiveness of this metric and the gains do come from diversity. (In Response 2/4)
  5. We test on a smaller model, LLaMA-3-8B-Instruct. DMAD also achieves the best performance. (In Response 3/4)
  6. We emphasize our novelty and differences compared with other methods leveraging diverse solutions. (In Response 3/4)
  7. We conduct experiments on the subset of a more challenging benchmark, MMLU [1], where DMAD also outperforms other methods. (In Response 4/4)
  8. We design 4 different stop criteria to determine when the debate ends to reduce excessive tokens. (In Response 4/4)

We also respond to other minor questions, and have added these supplements in the revised paper. We hope our responses can address your concerns. Looking forward to further communication with you.

[1] Measuring Massive Multitask Language Understanding. ICLR 2021.

评论

I appreciate the authors for their comprehensive response, especially in addressing my concerns about some of the design choices and the lack of more experiments. After consideration, I decided to raise my score to 6 and keep my rating neural. I still have concerns about the novelty of the work. The psychological perspective, while interesting, would, in my point of view, still not help enough to distinguish it well from other comparable works especially given ICLR standard. I do commend the soundness of the current experiments and writing and will leave ACs judgment to make the final decision.

评论

We are glad that our responses have addressed your concerns, and we are grateful for your commendation of the soundness of our experiments and writing. We sincerely appreciate and value your valuable opinions, which will help us to push the frontier in our future work.

评论
  • Weakness 3: Model capability of scaling.

We test MAD (All CoT), MAD (All SBP), MAD (All PoT), and DMAD with LLaMA-3-8B-Instruct on MATH. Results demonstrate DMAD is also effective on smaller models.

ModelsMethodsAlg.C&PGeom.Int. Alg.Num. Th.PreAlg.PreCalc.Avg.
LLaMA-3-8B-InstructMAD (All CoT)49.7±0.8219.3±1.6318.3±2.1615.3±0.8218.3±0.8246.7±1.6315.0±1.4126.1±0.42
LLaMA-3-8B-InstructMAD (All SBP)46.3±2.9420.7±4.9717.3±1.6315.3±0.8220.0±6.1644.3±5.3511.3±0.8225.3±1.94
LLaMA-3-8B-InstructMAD (All PoT)41.3±5.7220.0±2.8316.0±1.4113.7±4.5521.0±5.6639.7±4.9719.0±4.2424.4±1.62
LLaMA-3-8B-InstructDMAD47.7±0.8221.3±0.8221.7±0.8216.0±0.0021.3±5.7245.0±1.4115.0±1.4126.9±0.40

However, LLaMA-3-8B-Instruct using DMAD still gets lower accuracy (26.9%) than LLaMA-3-70B-Instruct using CoT (41.43%). We analyze that the upper bound of LLaMA-3-8B-Instruct using DMAD is still lower than 41.43%. We calculate diversitydiversity for {CoT, SBP, PoT} and get 0.3829. This means {CoT, SBP, PoT} can only correctly solve 38.29% of the problems, which is lower than 41.43%.

  • Weakness 5: Incremental rather than novel contribution.

We would like to emphasize our novelty and differences compared with other methods leveraging diverse solutions.

  1. We introduce a new concept of mental set and provide a specific definition and systematic analysis as above, which is inspired by the psychological theory of mental set.
  2. Self-Contrast just contrasts diverse reasoning solutions to get better performance than simple IO solutions. However, they don't compare their method with each distinct reasoning strategy. So it's not sure if using diverse solving perspectives all outperforms completely using one perspective among them. Meta-Reasoning Prompting dynamically chooses the best reasoning method to solve problems, while it only achieves a balanced performance between various reasoning methods and is often inferior to a specific one in most cases. Different from them, we conduct a more comprehensive analysis and prove that leveraging diverse reasoning methods can outperform all single methods with the same setting.
  3. Our experiment shows that DMAD has the potential to break the mental set of single reasoning methods. For example, for the same problem, all agents in MAD (All IO) constantly get wrong answers in all 5 rounds. While in DMAD, an agent thinking with IO correctly solves the problem after receiving messages from agents thinking from other perspectives in the 2nd round. This also holds for MAD (All CoT), MAD (All SBP), MAD (All PoT), MAD (All CCoT) and MAD (All DDCoT). We state this in Section 4.3.4 in the initial submission.
  • Question 2: Could the system dynamically adjust its reasoning methods during the debate process?

We assign each agent to think with a fixed reasoning method in DMAD. It's interesting to teach each agent to dynamically adjust its reasoning method. We mention that we leave it for future work in Appendix A "LIMITATIONS AND FUTURE WORK", "We can also dynamically guide agents to choose the method they deem appropriate independently. We leave this to future work and believe this may further improve the models’ reasoning performance." However, from another perspective, our experiments have proven that an agent thinking with a fixed method can break its mental set by getting insights from other diverse solutions. So it may not be necessary to design a dynamic adjustment strategy.

评论
  • Weakness 1: Resource efficiency analysis. Thanks for your advice. We report the token overhead and cost for each method in the experiment on MATH with GPT-4o-mini. DMAD balances the overhead of MAD (All CoT), MAD (All SBP), and MAD (All PoT), and can achieve the best performance than other methods.
MethodTokensCost ($)Accuracy
CoT-SC1,494,6920.780168.57
SBP-SC3,933,1341.887666.43
PoT-SC1,015,7050.344156.14
Self-Refine2,871,7641.027167.71
Self-Contrast6,159,0492.438962.14
MRP4,298,9262.029365.00
MAD-persona-D5,156,0171.274362.43
MAD-persona-E2,680,8710.682462.43
MAD (All CoT)4,445,0661.507769.86
MAD (All SBP)11,076,7923.541966.14
MAD (All PoT)3,113,7161.021567.76
DMAD (Ours)6,331,3162.044970.38
  • Weakness 2-3: Whether the gains are truly from method diversity or simply from the debate process itself?

  • Weakness 4: Limited implementation guidelines.

  • Question 1: Why choosing CoT, SBP, PoT? How about Least-to-Most prompting?

We initially chose divergent prompting strategies by intuition. CoT solves the problem step by step, but may struggle with problems involving complex theorems. SBP first extracts relevant principles and then solves the problem according to them, but may weaken the ability to solve problems step by step. PoT uses Python programs, but this is mostly suitable for problems involving numbers. Therefore, these methods are diverse and can complement each other. Instead, Least-to-Most (L2M) breaks down the problem into progressive sub-questions and answers them to get the final answer. It looks somewhat similar to CoT which gradually solves the problem in steps.

To offer a more quantitative analysis, we design a more objectively quantitative metric. Assuming we have KK candidate basic reasoning strategies Rii=1K\\{R_i\\}^K_{i=1} and want to select kk diverse ones, run each strategy NN times, and record the problems which RiR_{i} correctly solve at least once as PiP_{i} (Note that the definition of PiP_{i} here is different from PiP_{i} when introducing mental set). Note all problems on the measured dataset as PallP_{all}. We can define the diversity of the selected strategies {Rsi}i=1k\lbrace R_{s_i}\rbrace ^{k}_{i=1} as

diversity=Psii=1kP_all[0,1],si1,2,...K,sisjfor,,ij.diversity =\frac{\|\cup \\{{P_{s_i}\\}}^{k}_{i=1}\|}{\|P\_{all}\|}\in [0, 1], s_i \in \\{1, 2, ... \,K\\}, s_i \neq s_j \, for \\, \\, i \neq j.

This represents the proportion of total questions that the selected kk methods can answer correctly. The more diverse these methods are, the larger this proportion should be.

We run N=3N=3 times for each reasoning strategy in Rii=14=CoT,L2M,SBP,PoT\\{{R_i} \\}^{4}_{i=1}=\\{CoT, L2M, SBP, PoT\\} and select k=3k=3 strategies to calculate diversitydiversity, and test DMAD with different strategy groups on MATH with GPT-4o-mini. We can see diversitydiversity of {CoT, SBP, PoT} is the highest. Experiment results show that using the method group with larger diversitydiversity can get better results. This further proves gains are from method diversity, not the debate process.

ModelRs1{R_{s}}_{1}Rs2{R_{s}}_{2}Rs3{R_{s}}_{3}diversitydiversity
GPT-4o-miniCoTL2MSBP0.8471
GPT-4o-miniCoTL2MPoT0.8643
GPT-4o-miniCoTSBPPoT0.8657
GPT-4o-miniL2MSBPPoT0.8557
LLaMA-3-70B-InstructCoTL2MSBP0.6743
LLaMA-3-70B-InstructCoTL2MPoT0.6600
LLaMA-3-70B-InstructCoTSBPPoT0.7286
LLaMA-3-70B-InstructL2MSBPPoT0.6386
MethodAlg.C&PGeom.Int. Alg.Num. Th.PreAlg.PreCalc.Average
DMAD (CoT, L2M, SBP)88.7±0.8278.0±1.4154.7±4.5549.0±5.1082.7±0.8285.7±0.8237.7±3.2768.0±1.83
DMAD (CoT, L2M, PoT)91.7±1.6381.3±2.9454.0±1.4154.7±0.8282.3±0.8287.7±2.1639.0±2.8370.1±0.65
DMAD (CoT, SBP, PoT)91.7±1.6381.0±1.4157.3±2.1653.7±0.8282.7±0.8286.3±0.8240.0±1.4170.4±0.95
DMAD (L2M, SBP, PoT)87.0±1.4181.3±5.7254.7±3.2751.3±1.6380.7±4.5585.0±1.4138.0±3.7468.3±2.13
AC 元评审

This paper is a contribution to the literature on multi-agent debate. Reviewers were pleased that the paper was clearly written and showed clear improvements over appropriately chosen baselines. There were concerns about the novelty of the approach but no one thought they were insurmountable.

审稿人讨论附加意见

There was reasonably good engagement in the conversation. One reviewer increased their score after reading the author response.

最终决定

Accept (Poster)