PaperHub
6.6
/10
Poster4 位审稿人
最低3最高5标准差0.9
5
3
3
3
ICML 2025

MoRAgent: Parameter Efficient Agent Tuning with Mixture-of-Roles

OpenReviewPDF
提交: 2025-01-23更新: 2025-07-24

摘要

关键词
LLMsAgentMixture-of-Roles

评审与讨论

审稿意见
5

This work presents a novel framework to fine-tune LLMs to solve agent-specific task in a parameter-efficient manner. Specifically, the capabilities of agent are firstly decomposed into three roles. A relative fine-tuning framework MoR and a multi-role data generation pipeline are subsequently proposed to ensure that LLMs can correctly learn the capabilities of different roles.

给作者的问题

  1. Have you conducted additional experiments on larger models like 7B or 14B?
  2. In table 3, why qwen performs better after training when the base results are worse than llama on MATH benchmarks.
  3. Have the authors tried other role-decomposition strategies? For example, does using more or fewer roles have a greater impact on the results?

论据与证据

Yes, the author conducts several experiments with different size of models on different agent-specific tasks, and the results are convincing.

方法与评估标准

Yes. The author selects both agent-specific and mathematical benchmarks (which could also be solved in an agentic manner) to evaluate the proposed method.

理论论述

This work does not involve theoretical claims and proofs.

实验设计与分析

Yes. The three main tables have shown to be validate, the ablation studies also prove the correctness of the hyper-parameters and the design of the loss function.

补充材料

Yes, I have reviewed the supplementary material in the part of Appendix, which include the prompts used in completion and training process, and an example of the execution trajectory.

与现有文献的关系

The related scientific literature of the main contributions of the paper are already cited and studied. For example: the idea of role decomposition may relate to alpha-UMi[1], and the novel architecture may evolve from the work of OCTAVIUS[2] and MoLA[3]. [1] Small llms are weak tool learners: A multi-llm agent. [2] Octavius: Mitigating task interference in mllms via moe [3] Higher layers need more lora experts.

遗漏的重要参考文献

No, important related works are already cited and discussed in this work.

其他优缺点

Strengths:

  1. Other than directly training multi-models to learn different roles capabilities, the author introduces a novel architecture: Mixture-of-Roles, which comprises three specialized LoRA groups.
  2. The author has conducted comprehensive experiments on different benchmarks with different base models, the results show a significant improvement compared with the base models.
  3. The contributions of different parts in the loss function are studied through extensive experiments in the part of Ablation Studies. Weakness:
  4. To prove the effectiveness of the multi-role strategy, experiments with a single role is supposed to be conducted.
  5. When adapt to each new task, the work of data preparation and unification is essential, but seems time-consuming and labor-intensive.

其他意见或建议

Typos: Caption of Figure 8, ‘mathmatical’ -> ‘mathematical’

Suggestions: The code of the architecture and training process are suggested to be released.

作者回复

Dear Reviewer Mktg:

Grateful for your support and helpful review. All concerns and questions are meticulously responded.

1. Experiments with a single role is supposed to be conducted.

Thanks for the helpful suggestion. We supplement the experiment of a single lora without roles on BFCL leaderboard with Llama3.2-1B-Instruct.

MethodTrainable ParamsAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
Base-21.919.229.838.927.5
LoRA0.16B60.568.259.483.768.0 (+40.5)
MoR0.16B75.280.060.794.477.6 (+50.1)

With the same number of trainable parameters, due to the lack of decomposition capability, the accuracy improvement achieved by LoRA is much lower than that of MoR.

2. The work of data preparation and unification is essential, but seems time-consuming and labor-intensive.

Thanks for the kind concern. Actually, a large proportion of the raw training data lacks multi-roles content, such as the thought of reasoner. Therefore, it cannot be directly used for MoR training. The data processing pipeline proposed in the paper addresses this by generating high-quality multi-role data based on the raw dataset. However, as the reasoner paradigm like DeepSeek-R1 and OpenAI o3, etc., has become mainstream recently, there will be more agent related data with thought.

3. Additional experiments on larger models.

Thanks for the helpful suggestion. Due to the limited time of rebuttal, we only supplement the experiment of Llama3.1-8B-Instruct on BFCL.

MethodTrainable ParamsAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
Base-84.286.361.077.877.3
MoR0.59B88.689.280.595.188.4 (+11.1)

The experimental settings are consistent with section4.2 in our paper. From the results, with the introduction of 0.59B trainable parameters, we improve the average accuracy by 11.1%.

4. In table 3, why qwen performs better after training when the base results are worse than llama on MATH benchmarks.

The possible reason is that llama is a general model and qwen-coder is a code-specific model. Since we solve math problems by importing packages here, it is more in line with the style of qwen-coder, so the accuracy of qwen-coder can be improved higher in the post-training stage.

5. Using more or fewer roles have a greater impact on the results?

Thanks for the constructive suggestion, we supplement the experiments of more or fewer roles. The results are as follows.

ArchsTrainable ParamsAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
Base-21.919.229.838.927.5
2-roles0.11B60.666.348.379.563.7 (+36.2)
3-roles0.16B75.280.060.794.477.6 (+50.1)
4-roles0.21B70.874.257.690.373.2 (+45.7)

Our method integrates three core roles: Reasoner, Executor, and Summarizer. Through removing the Summarizer and introducing a Planner, we extend the roles to 2-roles and 4-roles. Experimental results demonstrate accuracy improvements across all variants compared to the baseline model. Notably, the 3-roles architecture achieves peak performance (accuracy improvement of 50.1% over baseline), whereas the 4-roles, despite containing more trainable parameters, shows diminishing returns (45.7% accuracy gain). This empirical evidence highlights that ​rational role definition outweigh mere parameter quantity expansion in achieving optimal model performance.

6. Typos.

Sincerely thanks for your detailed reviews. We will go through the entire paper again and fix potential typos in the revised version.

7. The code of the architecture and training process are suggested to be released.

Thanks, all our data and code will be open-sourced to contribute to the community.

审稿意见
3

This paper proposes multiple strategies to improve the efficiency of applying PEFT to agent. First, the capabilities necessary for the agent tasks are decomposed into three distinct roles: reasoner, executor, and summarizer. The Mixture-of-Roles (MoR) framework, which comprises three specialized LoRA groups, each designated to fulfill the three roles. To more reasonably allocate LoRAs to the input features, a rule-based and role-aware gate and learnable token-aware routers are designed. During the training process, auxiliary balance loss and orthogonal loss between LoRAs are further introduced for better optimization. Last but not least, a multi-role data generation pipeline is introduced.

给作者的问题

  1. Will the two gates lead to an increase in FLOPs or latency?
  2. What would be the result of having three separate LLMs to represent different roles?
  3. With the same parameters, what is the comparison between a large LoRA and multiple small LoRAs?

论据与证据

Yes

方法与评估标准

Yes, the proposed method is evaluated on StableToolBench、BFCL and Math problems.

理论论述

No proofs and theoretical claims.

实验设计与分析

Yes, I have checkd the soundness/validity of the experiments, including StableToolBench, BFCL leaderboard, GSM8K and MATH.

补充材料

Yes, I have reviewed the supplementary material in the appendix (Page11-Page15).

与现有文献的关系

Three key contributions in this paper:1) Decomposing the agent ability into three roles; 2) Each abilities are fulfilled by Mixture-of LoRAs; 3) A lot of work of preparing the CoT agent training data. Any relevant previous work has been discussed in the paper.

遗漏的重要参考文献

No.

其他优缺点

Strengths:

  • The idea of decomposing the ability of agent into three roles is interesting.
  • In my knowledge, this is the first work that apply Mixture-of-LoRAs to the agent tasks.
  • The composition of rule-based gate and role-aware gate is novel.
  • The experiments are conducted on StableToolBench, BFCL and Math, which are sufficient.

Weaknesses:

  • Some other work like [1] also introduces the idea of decomposition of agent ability, what is the difference between you?
  • The auxiliary balance loss is not novel.
  • Ablation experiments on the auxiliary balance loss and orthogonal loss weight coefficients are missing.
  • The preparation of experimental data seems to be a lot of work.

[1] Small llms are weak tool learners: A multi-llm agent

其他意见或建议

Typos: "available datasets To effectively" in Line435.

作者回复

Dear Reviewer MDgP:

We deeply appreciate your support and insightful feedback, detailed rebuttals to all queries are provided.

1. The difference between ours and α-UMi.

α-UMi decomposes the agent ability into planner,executor and summarizer. However, each role is implemented by a separate LLM, which results in a significant increase in computing resources. In contrast, we integrate the multiple agent capabilities into a novel parameter-efficient Mixture-of-Roles framework.

2. The auxiliary balance loss is not novel.

Thanks for the kind concern. Auxiliary balance loss is widely adopted in training MoE architectures. We borrow the idea of ​​balance loss and are the first to apply it to our agent work. The experimental results on various benchmark verify the effectiveness of the loss.

3. Ablation on the auxiliary balance loss and orthogonal loss weight coefficients.

Thanks for the helpful advice. we supplement the ablation studies on BFCL leaderboard with Llama3.2-1B-Instruct.

balance lossorthogonal lossAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
1e-31e-475.280.060.794.477.6
1e-41e-474.678.858.993.676.5
1e-31e-375.779.359.194.077.0
1e-41e-374.178.459.693.176.3

With appropiate coefficients of auxiliary balance loss and orthogonal loss, the accuracy may be further improved.

4. The preparation of experimental data seems to be a lot of work.

Actually, a large proportion of the raw training data lacks multi-role content, such as the thought of reasoner. Therefore, it cannot be directly used for MoR training. The data processing pipeline proposed in the paper addresses this by generating high-quality multi-role data based on the raw dataset. However, as the reasoner paradigm has become mainstream recently, there will be more agent related data with thought. All data and code will be open-sourced to contribute to the community.

5. Will the two gates lead to an increase in FLOPs or latency?

Thanks. Our gates include the rule-based role-aware gate and learnable token-aware gate. For the rule-based gate, a "role" tag is designed to indicate which role activates the subsequent tokens. Therefore, no FLOPs or latency is introduced. For each learnable token-aware gate, we implement it with only a single Linear layer, resulting in a negligible FLOPs and latency.

6. The result of having three separate LLMs to represent different roles.

This is an interesting question, and actually, that is precisely how α-UMi operates. We supplement this experiment on BFCL leaderboard with Llama3.2-1B-Instruct.

MethodTrainable ParamsAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
Base-21.919.229.838.927.5
MoR0.16B75.280.060.794.477.6 (+50.1)
3 LLMs3.72B79.383.269.594.881.7 (+54.2)

With more trainable parameters, three separate LLMs achieve higher accuracy, but the computational resources required for training and inference are much higher than ours.

7. The comparison between a large LoRA and multiple small LoRAs?

Sorry for not fully understanding what multiple small LoRAs meant, we guess it should be our proposed MoR. We supplement the experiment of one large LoRA with Llama3.2-1B-Instruct on BFCL.

MethodTrainable ParamsAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
Base-21.919.229.838.927.5
LoRA0.16B59.764.256.381.865.5 (+38.0)
MoR0.16B75.280.060.794.477.6 (+50.1)

The LoRA is trained on the same 90k multi-roles data. From the results, the proposed MoR can fully utilize the limited training parameters and achieve significantly better accuracy.

8. Typos.

Sincerely thanks for your detailed reviews. We will go through the entire paper again and fix potential typos in the revised version.

审稿意见
3

The paper introduces a novel parameter-efficient fine-tuning method to enhance LLMs for agent tasks, such as function-calling and mathematical reasoning. The authors propose three main strategies: (1) decomposing agent capabilities into three roles—reasoner, executor, and summarizer—based on the Reason+Action paradigm; (2) developing the MoR framework, which assigns specialized LoRA groups to each role, incorporating a rule-based role-aware gate and token-aware routers to manage role interactions; and (3) creating a multi-role data generation pipeline that enhances publicly available datasets with role-specific content and reliability verification. The method is evaluated on benchmarks like StableToolBench, BFCL, GSM8k, and MATH. The paper claims that MoRAgent achieves competitive performance with fewer trainable parameters than traditional methods.

update after rebuttal

Thank you for the response. I will keep my score.

给作者的问题

  1. How is the rule-aware gate implemented during training? Are there labels in the training data indicating which role should be active for each token? Clarification could resolve ambiguity in Section 3.2, strengthening the method’s reproducibility.
  2. How's the performance for standard LoRA? Comparing this method with standard LoRA can further consolidate the argument.
  3. I got a bit confused by section 3.2 and equation (6): "It should be noted that for the token at the same location, there is only one role that is non-zero" what exactly does this mean? I thought u is the hidden state, so each element in u should just be a value of the hidden state, not a token. I couldn't connect this with figure 2. Is each value in the hidden state activated by one role, or each token?

论据与证据

  • The paper claims that decomposing agent capabilities into three roles improves PEFT for agent tasks. The improvement from MoR is clear and comprehensive analysis is conducted to demonstrate that.
  • While we do see performance improvement on downstream tasks there is a lack of a direct ablation comparing MoRAgent with and without role decomposition (e.g., a single LoRA without roles), making it hard to isolate the decomposition’s specific contribution versus the multi-LoRA setup.
  • Another claim is that the MoR framework with a router works. But he rule-based role-aware gate’s implementation is vague. The paper states that “the next role to be activated is determined based on the output of the reasoner”, but lacks specifics on how this is operationalized during training, weakening the claim’s clarity.
  • The effect of multi-role data generation pipeline is unclear. No direct evidence compares MoRAgent’s performance with and without this pipeline versus raw data. Also it is unclear whether the performance is coming from MoR or the data pipeline.

方法与评估标准

The MoR framework decomposes agent tasks into reasoner, executor, and summarizer roles, Taligning with the Reason+Action paradigm cited in prior work. The proposed benchmarks are also reasonable and cover a wide variety of downstream tasks.

理论论述

The paper does not present theoretical claims or proofs. It is empirically driven.

实验设计与分析

  • The experimental designs are robust, I just have two concerns detailed above: the ablation between standard LoRA and MoR and the ablations for data pipeline. Both are to isolate the contribution of the MoR framework to make a stronger argument.

补充材料

N/A

与现有文献的关系

Builds on full-parameter fine-tuning works like ToolLLM (Qin et al., 2023) and AgentTuning (Zeng et al., 2023), which enhance LLMs for agent tasks and Inspired by Reason+Action (Yao et al., 2022) and multi-agent systems like α-UMi (Shen et al., 2024a). It is a novel extension on top of those works by making LLMs good tool users (therefore agent) while being resource efficient.

遗漏的重要参考文献

N/A

其他优缺点

  • Contribution is clear and simple: it addresses resource barriers in agent fine-tuning, relevant for deploying LLMs in resource-constrained settings.
  • However, the effectiveness of the method is unclear without the ablations mentioned earlier.

其他意见或建议

N/A

作者回复

Dear Reviewer rpMC:

We sincerely thanks for your support and meticulous review. The concerns and questions are answered as follows.

1. How is the rule-aware gate implemented during training? Are there labels in the training data indicating which role should be active for each token?

Yes! Your deduction is correct. Our training data includes a "role" tag to indicate which role activates the subsequent tokens. Take the Math data in Figure 3 in our paper as an example. The tokens marked as "content" are activated by the role marked as "role". It should be noted that when the role is user or observation, it is also activated by reasoner. Therefore, "system_prompt", "user_query", "thought and next role" and "code execution result" are activated by reasoner. "Python code" is activated by executor. "summarize and feed back to user" is activated by summarizer. We will further clarify this in the revised version.

2. The effect of multi-role data generation pipeline is unclear. It is unclear whether the performance is coming from MoR or the data pipeline. How's the performance for standard LoRA?

Thanks for the helpful suggestion. A large proportion of the raw training data lacks multi-role content, such as the thought of reasoner. Therefore, it cannot be directly used for MoR training. The data processing pipeline proposed in the paper addresses this by generating high-quality multi-role data based on the raw dataset. A comparative example is illustrated below.

Raw data:

{
    "role": "user"
    "content": "I want to watch a movie tonight. Can you suggest me some movies starring Leonardo DiCaprio?"
},
{
    "role": "assistant"
    "content": "[unused11]Action: search_movies\n Arguments: {"actor": "Leonardo DiCaprio"}[unused12]"
},
{
    "role": "observation"
    "content": "{"movies": ["Inception", "The Wolf of Wall Street", "Titanic", "The Revenant", "Shutter Island"]}"
},
{
    "role": "assistant",
    "content": " "Inception", "The Wolf of Wall Street", "Titanic", "The Revenant", "Shutter Island". "
}

Multi-role data after generation pipeline:

{
    "role": "user",
    "content": "I want to watch a movie tonight. Can you suggest me some movies starring Leonardo DiCaprio?"
},
{
    "role": "reasoner",
    "content": "To suggest movies starring Leonardo DiCaprio, we can use the "search_movies" tool with the input specifying the actor's name.\n Next: executor."
},
{
    "role": "executor",
    "content": "[unused11]Action: search_movies\n Arguments: {"actor": "Leonardo DiCaprio"}[unused12]"
},
{
    "role": "observation",
    "content": "{"movies": ["Inception", "The Wolf of Wall Street", "Titanic", "The Revenant", "Shutter Island"]}"
},
{
    "role": "reasoner",
    "content": "The function has been successfully executed and five movies starring Leonardo DiCaprio are returned, which can solve the user's query.\n Next: summarizer."
},
{
    "role": "summarizer",
    "content": "Sure, here are some movies starring Leonardo DiCaprio: "Inception", "The Wolf of Wall Street", "Titanic", "The Revenant", "Shutter Island". Enjoy your movies!"
}

All data and code will be open-sourced to contribute to the community. We supplement the experiments of LoRA in two settings: a single LoRA without roles (standard, trained on raw data) and a single LoRA with roles (multi-roles, trained on multi-role data).

MethodTrainable ParamsAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
Base-21.919.229.838.927.5
LoRA (standard)0.16B60.568.259.483.768.0 (+40.5)
LoRA (multi-roles)0.16B59.764.256.381.865.5 (+38.0)
MoR0.16B75.280.060.794.477.6 (+50.1)

Interestingly, the accuracy of LoRA (multi-roles) is lower than that of standard LoRA. The possible reason is that under limited training parameters, there is interference when a single LoRA learns the multi-roles datasets. In contrast, MoR can avoid interference by the architecture of mixture-of-roles to make full use of the limited training parameters and achieve higher accuracy.

3. Confused by equation(6).

Sorry for misleading. The size of u here can be represented by [sequence_length,hidden_dim] (for simplicity, we can omit the dimension of batch size). Without loss of generality, take the size of u in one layer equal to [4096,hidden_dim] as an example. Further example, [:1024, :] of u is processed by reasoner, [1024:1536, :] of u is processed by executor, [1536:2560, :] of u is processed by reasoner, [2560:3072, :] of u is processed by executor, [3072:3584, :] of u is processed by reasoner, and [3584:, :] of u is processed by summarizer. We will further clarify this in the revised version.

审稿意见
3

This paper explores parameter-efficient fine-tuning (PEFT) methodologies for large language model (LLM)-based agent tasks, an area that remains largely unexplored. The authors propose three key strategies:

  1. Role Decomposition: Inspired by the Reason+Action paradigm, the authors decompose agent capabilities into three distinct roles—reasoner, executor, and summarizer. The reasoner interprets user queries and determines the next step based on execution trajectory. The executor identifies and invokes appropriate functions with the correct parameters. The summarizer distills and conveys information back to the user.

  2. Mixture-of-Roles (MoR) Framework: The authors introduce a framework with three specialized Low-Rank Adaptation (LoRA) modules, each dedicated to a specific role. These modules collaboratively perform the agent task while maintaining parameter efficiency.

  3. Multi-Role Data Generation Pipeline: A novel data generation pipeline is designed using publicly available datasets. It incorporates role-specific content completion and reliability verification to support fine-tuning the MoR framework.

The paper presents extensive experiments and ablation studies on various LLMs and agent benchmarks, demonstrating the effectiveness of the proposed approach.

给作者的问题

  1. What's the gain by SFT on the same datasets, w/o lora? and other PEFT methods?

  2. Could you validate your approach on more agent structures? such as Reflection, AutoGen?

  3. Could you validate your approach on more agent-related datasets? The training data is tool-use and the validation set is also tool-using (except for math), how about generalization ability on other agent tasks, maybe refer to Agentgym.

论据与证据

yes

方法与评估标准

yes

理论论述

yes

实验设计与分析

yes

补充材料

yes

与现有文献的关系

good

遗漏的重要参考文献

no

其他优缺点

Strengths:

  1. Novel PEFT Approach for Agents: The paper addresses a relatively unexplored area—parameter-efficient fine-tuning (PEFT) for agent tasks—by introducing a structured role-based approach.

  2. Clear Role Decomposition: The division of agent capabilities into reasoner, executor, and summarizer aligns well with the Reason+Action paradigm, making the framework interpretable and modular.

  3. Efficient Fine-Tuning via LoRA: The use of Mixture-of-Roles (MoR) with specialized LoRA modules enables efficient adaptation of LLMs without full model fine-tuning, reducing computational overhead.

Weaknesses:

  1. Limited Comparison with Other PEFT Methods: While the paper focuses on its novel approach, it lacks a direct comparison with other existing PEFT techniques that might be adapted for agent tasks.

  2. Scalability and Generalization: The approach is tailored to a specific role-based agent structure, and its effectiveness for more complex or diverse agent architectures remains unclear.

  3. Data Dependence: The proposed multi-role data generation pipeline relies on publicly available datasets, but its adaptability to real-world or unseen tasks is not extensively analyzed.

  4. Limited experiments about the ablation. It is unclear how much gains contributed from the SFT on the such diverse datasets.

其他意见或建议

see Question

作者回复

Dear Reviewer QZb4:

Sincerely thank you for your constructive review, the concerns and questions are answered in detail.

1. What's the gain by SFT on the same datasets, w/o lora? and other PEFT methods?

Thanks for the helpful suggestion. Based on the same multi-roles dataset, we supplement the experimental results of model Llama3.2-1B-Instruct on BFCL leaderboard. The results are as follows.

MethodTrainable ParamsAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
Base-21.919.229.838.927.5
SFT1.24B72.377.661.592.676.0 (+48.5)
LoRA0.16B59.764.256.381.865.5 (+38.0)
DoRA0.16B61.265.758.482.066.8 (+39.3)
Ours0.16B75.280.060.794.477.6 (+50.1)

From the results, SFT exhibits superior accuracy compared to PEFT methods (LoRA and DoRA), which can be attributed to its more trainable parameters, achieving an average accuracy 10.5% higher than LoRA. Notably, DoRA introduces an advanced scheme by decomposing pretrained weight matrices into magnitude vectors (m) and directional matrices (V), where LoRA is applied specifically to V while m is trained separately. This architectural innovation allows DoRA to surpass LoRA slightly in accuracy. Crucially, our proposed method achieves statistically significant performance improvements through two key innovations: 1) a more rational capacity decomposition strategy, and 2) a novel Mixture-of-Roles framework enabling dynamic interaction between decomposed modules. These enhancements collectively contribute to our method's marked accuracy superiority over SFT and PEFT methods.

2. Validating on more agent structures, such as Reflection, AutoGen.

Thanks for the valuable suggestion. Our method is not constrained by agent architectures. For example, the ​framework of Reflection[1] comprises three modules: ​Actor, ​Self-Reflection, and ​Evaluator. Similarly, in ​AutoGen's[2] application scenarios (Figure 3), multi-agent coding involves modules like ​Commander, ​Writer, and ​Safeguard. While each module in these frameworks operates as an independent LLM, our approach integrates multiple independent LLMs into a ​parameter-efficient Mixture-of-Roles architecture.

Due to the limited time during rebuttal, directly extending our method to these application scenarios proved challenging. Therefore, we modified the agent architecture by adding or removing specific roles based on the existing ​BFCL multi-roles training dataset. The results are as follows.

ArchsTrainable ParamsAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
Base-21.919.229.838.927.5
2-roles0.11B60.666.348.379.563.7 (+36.2)
3-roles0.16B75.280.060.794.477.6 (+50.1)
4-roles0.21B70.874.257.690.373.2 (+45.7)

Our method integrates three core roles: Reasoner, Executor, and Summarizer. Through architectural modifications — specifically removing the Summarizer and introducing a Planner — we extend the framework to 2-roles and 4-roles. Experimental results demonstrate accuracy improvements across all variants compared to the baseline model. Notably, the 3-roles architecture achieves peak performance (accuracy improvement of 50.1% over baseline), whereas the 4-roles, despite containing more trainable parameters, shows diminishing returns (45.7% accuracy gain). This empirical evidence highlights that ​rational role definition and architecture design outweigh mere parameter quantity expansion in achieving optimal model performance.

[1] Reflexion: Language Agents with Verbal Reinforcement Learning

[2] AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation

3. Validating on more agent-related datasets. How about generalization ability on other agent tasks, maybe refer to AgentGym.

Thanks. In our experiments, we conducted evaluations across multiple datasets including StableToolBench, BFCL, GSM8K, and MATH. In fact, the application scenarios of these datasets involve API calls, Java programming, Python programming, JavaScript programming, and mathematical problem-solving, etc., not just tool use. AgentGym is an impressive work, however, limited by the time of rebuttal, it is difficult for us to reproduce AgentGym in a short time and migrate our methods to it. However, we supplement the experiments of Llama3.2-1B-Instruct on BFCL with varying amounts of training data.

Train DataAST (Non-live)Exec (Non-live)AST (Live)Relevance (Live)AVG
021.919.229.838.927.5
1k49.544.646.379.154.9 (+27.4)
5k55.351.850.682.160.0 (+32.5)
10k58.857.452.785.963.7 (+36.2)
50k70.474.956.591.773.4 (+45.9)
90k75.280.060.794.477.6 (+50.1)

From the results, even with only 1k training samples, we still achieves a 27.4% improvement in average accuracy, demonstrating its strong generalization capability. As the training data volume increases, the accuracy further improves accordingly.

最终决定

This paper present a parameter-efficient fine-tuning method (mixture of roles framework) for agent tasks and received consistently positive feedback from all four reviewers, i.e., three Weak Accept, one Strong Accept. The authors provided additional experiments and clarifications during the rebuttal to address the raised concerns, supporting a recommendation for acceptance.