PaperHub
5.8
/10
Poster3 位审稿人
最低3最高4标准差0.5
4
4
3
4.0
置信度
创新性2.7
质量2.7
清晰度2.7
重要性2.3
NeurIPS 2025

Incentivizing Reasoning for Advanced Instruction-Following of Large Language Models

OpenReviewPDF
提交: 2025-04-11更新: 2025-10-29

摘要

关键词
instruction followinglarge language modelreasoning

评审与讨论

审稿意见
4

In this paper authors propose a method to boost LLM performance on complex instruction-following tasks by constructing a set of complex instructions that was used for training different language models with Reinforcement Learning. Data is obtained by using self-instruct technique from publicly accessible instruction tuning datasets and RL verification is done with both rules and LLM-as-a-Judge. The method shows promising results for smaller LMs (1.5B) on complex instruction-following benchmarks.

优缺点分析

Strengths

  • Comprehensive experiments - multiple model families and model sizes (1.5B, 7B, 8B).
  • Extensive evaluations and good baselines.
  • Smaller models (1.5B parameters) with RL substantially outperform baselines.
  • Open-source models and datasets were used.

Weaknesses

  • Results are not convincing:
    • Table1: all larger models (7B+) show either negligible improvement over I/O or even worse performance (Llama3.1-8B) compared to baselines.
    • Table1: for Qwen 2.5-1.5B SFT gives already substantial boost.
    • Table 2: considering different compute spent, not clear if there's an advantage for ComplexBench vs baselines.
  • Comparison is better done on the aligned compute since it's not clear what training/test-time compute was spent, we can be sure that the proposed approach has any advantage.

问题

  • Regarding Table 2:
    • Can authors please provide more details on oracle decomposition?
    • What was the compute spent for each method?
  • What LLM was used for self-evolving instruction generation?
  • Effect of Behavior Cloning. From Table 1, e.g. Qwen2.5-1.5B-Instruct (cold-start) has superior performance after RL over the baselines while DeepSeek-Qwen7B (warm-start) has a marginal performance over I/O. At the same time LLaMA3.1-8B-Instruct (cold-start) performance deteriorates drastically after RL. How this corresponds to "Without proper guidance, cold-start LLMs can only repeat their inherent shallow thinking and consistently receive negative feedback"?

局限性

yes

最终评判理由

Thank you for providing additional evaluation data and more clarifications. My main concerns (same compute for train/test, 1.5B SFT being good) are resolved.

格式问题

No concerns.

作者回复

Q1: Why LLaMA3.1-8B collapses?

Existing studies use Qwen for incentivizing reasoning, and there exists few studies that report successful LLaMA performance gains. Why? It does not work well on LLaMA and training easily collapses halfway [97]. As we reported, the training logs (Fig. 55) show a collapsing at step 1K that pushed πθπ_θ into a degenerate distribution.

The collapsed LLaMA exhibits the symptoms:

  1. it keeps repeating reasoning contents without jumping to final answer;
  2. it fails to output valid response but only the assistant-like call-closing statements in Chinese.

We believe such phenomenon is related to:

  • limited multi-language capability of LLaMA (especially in Chinese comprehension and generation)
  • insufficient basic knowledge of step-by-step reasoning (lack of pretraining on massive mathematic corpus).

In this case, we should interpret the results of LLaMA3.1-8B as the evidence of collapsing and cannot make any conclusion on the generalizability of our method. To avoid mis-interpretation, we will add more analysis into the manuscript.

Q2: 1.5B is already good w/ SFT, why we bother w/ RL?

SFT might be a good starting point, but it does not promise generalizability.

  • Our method outperforms SFT on all 1.5 B models (+3.68 pp on Qwen1.5B, +6.85 pp on DeepSeek1.5B, +6.85 on DeepscaleR 1.5B).
  • Our method outperforms SFT on all 7~8B models where SFT can barely boosts LLMs consistently on all benchmarks.

To improve the generalizability of SFT, one must prepare millions of instructions for injecting a broad range of reasoning patterns into existing instructed models. With RLVR, online rollout is performed for preference alignment as long as the complex instructions and their verifications are prepared. It encourages LLMs to self-develop reasoning that enjoys a higher level of generalization.

Q3: Comparison with SOTAs on training and testing-time compute?

We report the training and testing-time compute in the table below (Qwen2.5-7B-Instruct). We followed the official implementation (if publicly released) and used the same instruction datasets for fair comparison on ComplexBench.

Training ComputeTest-Time Computeδ\delta
Method#SFT samples#SFT epochsRLHF algo.#RLHF samples#RLHF epochsAvg. # steps or samplingAvg. # reasoning tokensAvg. # answer tokensAvg. gains over I/O
I/O----------1--33274.47
CoT----------180419-2.46
SDC----------179372+2.06
SFT13K10------1--361-0.23
OD [11]----------1--407+1.79
SC [98]----------10--3302-0.70
SR [99]----------1.9--877-0.46
CNFR [14]66K4DPO63K11--308-10.96
FC [13]12K2DPO12K21--261-2.50
Ours----GRPO26K31299349+2.93

We observe that:

  • Most training-free methods use manually designed workflows (e.g., alternating reasoning and response drafting, iterative refinement, or consistency-based sampling) to increase test-time compute. However, as discussed in the manuscript (Fig. 1, lines 36–48), naive CoT performs poorly on instructed models with complex instructions. Shallow reasoning leads LLMs to focus only partially on the original instruction and overlook sub-instruction structure. Self-refinement [99] also fails to improve results, as LLMs tend to be overconfident and stop refining after just 2 turns on average (max 6).

  • The self-consistency method (10 samples) uses ten times more tokens but shows only a slight performance drop, indicating that sampling multiple responses does not improve performance on ComplexBench. Since complex instructions involve multiple constraints and nested structures, deep reasoning is needed for comprehension, not just more sampling.

  • The SDC method, by separating reasoning from answer generation, improves performance. Prompting the LLM to generate only reasoning first allows it to better analyze sub-instructions and constraints. All reasoning is then combined with the original instruction for answer generation.

  • Oracle decomposition tests if LLMs perform better with simplified instructions (removing compositional structures). Results show slightly longer responses and better scores, suggesting that redundant constraints in complex instructions can confuse LLMs. However, such decomposition requires human annotation and is not practical in real use.

  • Most training-required methods use SFT+DPO to improve LLMs on complex instructions, requiring ground-truth SFT responses and carefully selected DPO pairs [13][14]. In contrast, our GRPO-based RLVR does not need contrastive pair design and benefits from online rollout for easier pipeline setup of optimization.

  • Compared to CNFR [14] and FC [13], our method increases test-time compute by an average of 299 tokens due to reasoning, which we believe is crucial for LLM generalizability and worth the modest cost. For training compute, CNFR [14] uses the most SFT examples, while SFT often mixes general and complex samples to prevent forgetting [14]. Thus, training compute does not guarantee better performance if the SFT dataset lacks diversity in instruction constraints and structures.

We will add the above comparison on training and test-time compute into the manuscript. More explanations will be added into the discussion section.

Q4: What is oracle decomposition?

We follow ComplexBench [11] to implement oracle decomposition. It is designed to test whether the decomposition of instructions and executing them step-bt-step can improve the performance of LLMs.

  • And: remains as it is;
  • Chain: is divided into sequential tasks where each task corresponds to one sub-instruction, and then executed one by one each time;
  • Selection: is divided into selection and execution branches where only the true valid sub-instructions are kept;
  • Nested: is processed as a combination of the above.

We compare the performance of LLMs between the vanilla I/O (excution of the entire complex instruction in one step) and the oracle decomposition (execution of the decomposed instructions step-by-step).

In practice, there exists no ground-truth decomposition and therefore the oracle decomposition can be interpreted as how the cumulative errors would be if multi-round interactions are performed for LLMs to sequentially respond to each sub-instruction.

Q5: Which LLMs are involved in evolving of instructions?

Three LLMs (see Table 7):

  • Seed Instruction Selection: TagLM [87]
  • Evolving with Rules and Constraints: Qwen2.5-72B-Instruct [95]
  • Response Gneration and Quality Check: DeepSeek R1 [18]

The choice of LLMs in different stages can be determined by the available LLMs at hand. It is free to choose any competitive open-source or proprietary LLMs.

We also agree with the reviewer oTH1 that the expression of self-evolving is not precise because we use larger LLMs for instruction evolving rather than the LLMs used in training. Therefore, we would like to replace the term self-evolving with LLM-based instruction evolving.

Q6: How to interpret behavior cloning?

We would like to resolve any mis-understanding about line 289.

First, the experimental results of the ablation on behavior cloning (BC) are provided in Table 3. Both line 7 (w/o BC) vs. line 9 (w/ BC) and line 8 (w/o BC) vs line 5 (w/ BC) confirm the necessity of applying behavior cloning.

Second, with respect to Without proper guidance, cold-start LLMs can only repeat their inherent shallow thinking and consistently receive negative feedback, we are saying that the instructed models (i.e., non-reasoning model) are not pretrained on reasoning corpus in general domains and therefore they exhibit shallow reasoning during RLVR. In this case, we perform behavior cloning on instructed models so that 1) the adherence to the expected format allows successful parsing and grading of answers at an early stage; 2) the deep reasoning on complex instructions distilled from DeepSeek R1 can be imitated; 3) the potential reward hacking can be mitigated to prevent over-optimization.

Third, it is unfair to compare the cold-start and warm-start models with different model sizes. Under the same model size (1.5B), both our cold-start instruct model (Qwen2.5-1.5B-Instruct) and warm-start reasoning models (DeepscaleR-1.5B and DeepSeek-Qwen1.5B) exhibit significant superiority over baselines.

Fourth, with respect to the 7B-scale models, althought their performance gains are smaller than 1.5B models, the consistent improvement across cold-start and warm start models are non-trivial. One could easily verify that previous studies are validated on very few benchmarks (mostly IFEval, FollowBench or InfoBench) and their generalizability is not fully examined. Our improvement on the extensive SEVEN benchmarks is hard to achieve, thanks to the powerful tool of RLVR which allows LLMs to generalize reasoning-inspired instruction following rather than overfitting.

Last but not least, we would like to explain that for larger models, the current dataset size (13K) might not be enough. The TULU 3 authors use 15K RLVR instructions for the IFEval-style instructions and their optimized LLMs exhibit limited generalizability on OOD constraints. In contrast, we only construct 13K instructions and achieve generalizability on brand-new constraints from IFBench (see our new OOD evaluation in our response to reviewer oTH1). Therefore, we believe with an increase of dataset size, 7B-scale models are expected to achieve superior performance according to the scaling law (Kimi-arxiv 2501.12599 & OpenThoughts-arxiv 2506.04178). In the future, we plan to construct mixed RLVR dataset with size over 100K for scaling up.

More explanations will be added to make it clearer.

评论

Dear Reviewer U6ie,

Thank you sincerely for taking the time to review our work. We sincerely appreciate your thoughtful comments and evaluations. In our response, we have addressed your concerns point by point. If you have any further questions or feedback, we would be happy to continue the discussion.

Best regards,

Authors

评论

Thank you for providing clarifications and additional evaluations. Most of my concerns have been resolved, so I will raise the score to 4.

评论

We sincerely thank the reviewer for the positive feedback. We are glad that our responses addressed the raised concerns. The relevant clarifications are incorporated into the the manuscript to continue improving the quality of the manuscript.

审稿意见
4

With a focus on improving language models' ability of following constrained instructions, this work presents a dataset construction strategy, and an appropriate task-specific reward function to be used for training models with RL (particularly GRPO). The dataset is an augmentation of existing instruction tuning datasets (e.g., WildChat) with additional constraints and associated verification from either unit tests of LLM judgments. The constraints used in the training dataset overlap with those used in the evaluation datasets, which include IFEval, CELLO, FollowBench, etc. Experiments with various <8B parameter base models show that the training with GRPO on the constructed dataset improves performance on the evaluation sets, more so than CoT prompting and SFT.

优缺点分析

Strengths:

  • The proposed approach is straightforward.
  • It is nice to see that it improves performance over expected baselines with a range of base models.

Weaknesses:

Main concerns:

  • The taxonomy of constraints used in the training dataset are the same as those in the evaluation datasets. This leaves open the question of generalizability of the proposed approach to new unseen constraints. It would be helpful to hold out some constraints for an OOD evaluation to measure generalizability.
  • The proposed idea is not new. Using RL with verifiable rewards for the problem of constrained instruction following was done for training Tulu 3 (Lambert et al., 2024; https://arxiv.org/abs/2411.15124; Section 6). Related to the point above, they show that this approach does not generalize to new constraints. It would be interesting to see if the same trend holds in the current work too. Also, it is worth discussing if and how the current work differs from that work.

Additional concerns:

The writing in the paper can generally be improved. Documenting specific issues here:

  • The term "self-evolving" occurs throughout the paper, but it is not clear what this means. Section 3.2 says this method is an adaptation of Self-Instruct (SI), but SI uses the same model to generate the instructions and the responses, but this approach augments existing instructions (so it is unclear what the "self" refers too), and as far as I can tell, there is no "evolution" of instructions in this approach.
  • In the Introduction of the paper, the problem itself can be better defined as instruction following with complex verifiable constraints, along with examples of such instructions. To someone who is not already familiar with the problem, it may not be clear what this paper is about.
  • The second paragraph says prior approaches are all task-specific and are prone to over-fitting. This claim needs a citation.

问题

  • How well does the approach perform on unseen constraints?
  • How is the approach different from RLVR targeting IFEval in Lamber et al., 2024?
  • What do the authors mean by "self-evolving"?

局限性

Yes

最终评判理由

I find the additional discussion and the new results to be informative, and they address my two main concerns well.

格式问题

None

作者回复

Q1: Does the proposed method generalize on OOD Constraints?

We find it quite interesting to investigate the generalizability of the proposed method in handling constraints that are brand-new and human-annotated. We follow the reviewer's suggestion and carefully study the paper of Tulu 3 (Lambert et al.-arxiv 2411.15124) for experimental settings. We use their most recently released benchmark IFBench (Pyatkin et al.-arxiv 2507.02833) as OOD evaluation dataset because:

  1. it is proposed by AI2 that releases Tulu 3;
  2. it is an extension to the IFEval-OOD dataset which contains 58 new, diverse, and challenging constraints that are different from existing benchmarks like IFEval;
  3. it is released in July and there exists no possibility of data contamination with respect to our work.

Based on their official implementation, our evaluation results are shown below:

Modelprompt-level strictinstruction-level strictprompt-level looseinstruction-level looseavg.
Qwen2.5-1.5B15.6417.0118.3620.2917.82
+Ours17.6819.420.0622.6819.95(+2.13)
DS-Qwen-1.5B8.8011.6412.9215.8212.29
+Ours12.9214.0215.6416.7114.82(+2.53)
DeepScaler-1.5B11.2212.2315.317.9114.16
+Ours12.5814.0217.0018.8015.60(+1.44)
Qwen2.5-7B28.2329.8531.6333.4330.78
+Ours29.8230.7632.2735.4332.07(+1.29)
Ministral8B16.6617.3123.1224.4720.39
+Ours20.7423.8828.2331.9426.19(+5.77)
DS-Qwen-7B13.614.6219.7222.0817.50
+Ours20.0622.3825.1727.4623.77(+6.27)

Our method achieves consistent improvement on the instruction following capabilities across model sizes and families, demonstrating the generalizability on new constraints. It suggests that the proposed method does NOT overfit the constraints of IFEval.

The Tulu 3 paper reports that their optimized LLMs overfit IFEval and exhibit performance degradation compared with models before RLVR (Table 30 in their paper). This contradicts with our findings, which also highlights our advantages over existing RLVR methods.

We believe the generalizability of RLVR for solving complex instructions lies in three key aspects:

  • The diversity of constraints. Tulu 3 fully exploits the constraints in IFEval to augment existing instruction datasets (e.g., Wildchat). We do not know their exact construction details, but we are aware that the distribution of constraints in IFEval benchmark is severely unbalanced. In this case, as disclosed in our detailed construction process (Sec. A.3), we refer to the atomic constraint definitions of IFEval (Fig. 7) and generate the evolved instructions with multiple, diversified constraints. We do NOT overfit the distribution of constraints in the testing set to sacrifice the generalizability of the proposed method.
  • The diversity of compositional structures. Tulu 3 simply considers the IFEval constraints that are merely code verifiable. However, we consider semantic constraints by introducing a LLM for judgement. In addition, Tulu 3 only considers the "AND" composition for multiple constraints. On the contrary, we consider all structures including "AND", "SELECTION", "CHAIN", and "NESTED". Such semantic constraints verified by LLM-as-a-Judge also benefit our method in reducing the over-optimization problem (see Fig. 26, 27 in Tulu 3 paper).
  • The power of deep reasoning. Tulu 3 uses RLVR to directly optimizes the responses without deliberate reasoning. In contrast, we focus on improving "reasoning" of LLMs for complex instructions, which is a "pre-requisite" to the final answers. It is noted that the reasoning is a very powerful tool for generalizability, which has been confirmed in our study (see Sec. A.7.2).

We will add the above Table and its explanations into the manuscript as generalization study.

Q2: What are the differences between the proposed method and Tulu3?

To improve clarity, we will add Tulu 3 in related work and highlight our differences.

  • Problem Definition. We aim at solving a broad range of complex instructions with various constraints and their compositions via reasoning. We are dealing with the shallow reasoning problem of existing LLMs under this scenario. Tulu 3 focuses only on aligning model responses with IFEval-style instructions and therefore exhibits limited generalizability.
  • Constraint Definition and Structure. We follow CFBench [8] and ComplexBench [11] respectively for the constraint categories (e.g., word, lexical, and semantics) and composition types (e.g., And, Selection, Chain, Nested). Tulu3 is limited in code-verifiable constraints with "and" structures (all constraints have to be satisfied simultaneously).
  • Reward modeling. Both code-execution and LLM-as-a-Judge are involved for verification. This greatly expands the tasks that can be handled in our method. It is noted that no previous studies ever evaluate the capability of solving complex instructions on such a scale where seven comprehensive benchmarks are involved. As pointed out by Tulu 3, it is very challenging to generally improve the instruction following of LLMs as over-optimization could occur during RL. Our semantic constraints with LLM-as-a-Judge verifications are complementary to the lexical and word constraints in Tulu 3, which prevents reward hacking.
  • RLVR algorithm. The Tulu 3 is designed to directly optimize the response via the vanilla PPO algorithm without deliberate reasoning. However, we are aimed at incentivizing reasoning for solving complex instructions. We have to point out that the reasoning is NOT natural under our scenarios as it is under mathsmatic problems. Since the LLMs are prone to shallow reasoning and tend to muddle with a few sentences that ignore the constraints and their structures. In this case, we design the filtering mechniasm that only selects responses with deep, true reasoning. During rollout stage, we generate responses with and without reasoning at the same time. If the responses with reasoning are scored lower than those without reasoning, we believe the reasoning content is negatively associated with the answer and should be discarded to encourage the beneficial ones.

Q3: What is self-evolving exactly?

First, we would like to explain that we DO NOT simply augment existing instructions as Tulu 3. We DO NOT follow their simple technique of adding verifiable constraints into the original instructions. Tulu 3 only involves IFEval-style constraints and ignores semantic ones. They do not consider the structures of constraint organization at all. In addition to the naive "AND" composition, we build "SELECTION", "CHAIN", and "NESTED" compositions, which are missing in existing instruction datasets such as WildChat. Therefore, evolution of instructions is performed to incorporate multiple sub-instructions.

Second, with respect to "self-evolving", we would like to emphasize that these complex instructions are generated by LLMs and are evolved from the ones in datasets like WildChat. All the prompts for evolving instructions have been described in Sec. A.3.

Third, we agree with the reviewer that that "self-" is not accurate since we use large LLMs (see Table 7) for generation of the instructions. In this case, we will modify the "self-instruct" expression with "LLM-based evolving" for clarity.

Q4: Is the problem definition precise enough?

The current problem definition follows ComplexBench [11] and the terms might not be straightforward for people not familiar with the concept.

We agree with the reviewer that the introduction will be polished with clearer problem definition.

Q5: Can we support the claim on generalization with a citation?

We will add TULU 3 for citation as the reviewer mentions the generalizability of their work is limited.

prior approaches are prone to over-fitting constraints in the training set and fail to generalize to the unseen ones [Lambert et al., 2024, Tulu 3]"
评论

Thank you very much for your detailed response. It is nice to see improvements due to your approach on IFBench. These results indeed address my concerns regarding generalizability to new constraints. Can you please also include a comparison with the baselines (CoT, SDC, and SFT) on IFBench? I am also convinced by your discussion on comparing your approach with that from the Tulu 3 work.

评论

We sincerely thank you for helping us polishing the manuscript. Such suggestions not only enrich the investigation of related work (e.g., TULU 3) but also help us reflect on what we truly benefit from the reasoning formulated via RLVR (e.g., generalizability).

We provide the full comparison with baselines below.

ModelMethodprompt-level strictinstruction-level strictprompt-level looseinstruction-level looseavg.
Qwen2.5-1.5BI/O15.6417.0118.3620.2917.82
Qwen2.5-1.5BCoT13.6015.5215.9817.6115.67(-2.14)
Qwen2.5-1.5BSDC15.9817.6117.6819.7017.74(-0.08)
Qwen2.5-1.5BSFT16.3217.6119.3820.8918.55(+0.73)
Qwen2.5-1.5BOurs17.6819.420.0622.6819.95(+2.13)
DS-Qwen-1.5BI/O8.8011.6412.9215.8212.29
DS-Qwen-1.5BSFT12.1313.6914.1615.9813.99(+1.69)
DS-Qwen-1.5BOurs12.9214.0215.6416.7114.82(+2.53)
DeepScaler-1.5BI/O11.2212.2315.317.9114.16
DeepScaler-1.5BSFT12.7113.9415.0216.7414.60(+0.44)
DeepScaler-1.5BOurs12.5814.0217.0018.8015.60(+1.44)
Qwen2.5-7BI/O28.2329.8531.6333.4330.78
Qwen2.5-7BCoT27.8930.4430.9533.7330.75(-0.03)
Qwen2.5-7BSDC26.8729.2532.3134.6230.76(-0.02)
Qwen2.5-7BSFT23.1226.5728.5732.5427.70(-3.08)
Qwen2.5-7BOurs29.8230.7632.2735.4332.07(+1.29)
Ministral8BI/O16.6617.3123.1224.4720.39
Ministral8BCoT15.3014.9229.5931.3422.78(+2.39)
Ministral8BSDC18.3618.8023.8024.7721.43(+1.04)
Ministral8BSFT12.2413.7316.3219.415.42(-4.96)
Ministral8BOurs20.7423.8828.2331.9426.19(+5.77)
DS-Qwen-7BI/O13.614.6219.7222.0817.50
DS-Qwen-7BSFT17.3418.8021.0822.6819.97(+2.47)
DS-Qwen-7BOurs20.0622.3825.1727.4623.77(+6.27)

From the full comparison between the proposed RLVR method with existing baselines (e.g., CoT, SDC, and SFT), we report the following new findings:

1) For three out of four instructed (i.e., non-reasoning) models, the vanilla CoT does not bring gains but performance drop on the new constraints of IFBench, which is in line with our findings in Table 1. The shallow reasoning nature of instructed models often cause misinterpretation of the original instructions and neglect of certain constraints. They only briefly summarize the original instruction without truly analyze the rules to be obeyed for generation.
2) The SDC decouples the thinking and execution via two separate prompting. Such isolation allows the models to review the original instruction again, where the previous shallow reasoning only acts as a part of the context. Therefore, the SDC performs better than CoT but it is still prone to the unperfect reasoning.
3) For reasoning models, SFT brings a slight improvement as they learn the distilled reasoning patterns from a stronger reasoning model (e.g., DeepSeek R1). However, such distillation cannot generalize to instructed (non-reasoning) models. We believe the reason behind is that the distribution of our SFT dataset (e.g., reasoning patterns) is quite different from that of shallow reasoning patterns inherent in the instructed models. 
4) **As demonstrated by DeepSeek R1 [arxiv 2501.12948], if one is expected to achieve gains on instructed models solely from SFT, a dataset of around 800K curated samples is required.** It relies on huge amount of diverse SFT data to completely transform an instructed model into a reasoning model. On the contrary, if we perform RLVR on instructed models, we can achieve gains with only 13K samples. Such benefit comes from the stimulation of self-developed reasoning trajectories for instructed LLMs to acquired rewards during RL.

The OOD study above is added into our manuscript together with the related discussions. If there are still any remaining concerns, we would greatly appreciate the opportunity to clarify them.

评论

Dear Reviewer oTH1,

We sincerely thank you for the review and your constructive suggestions.

In our response, we have addressed your concerns point by point, especially around the new findings of generalization on OOD constraints (which are conducted on the IFBench released by TULU 3 team). More explanations and discussions are incorporated into the manuscript for consistent quality improvement.

As the discussion period is nearing its end, we would like to kindly check whether there are further questions. We would be happy to continue the discussion.

Best regards, Authors

审稿意见
3

The article introduces UTU 1.0, a systematic method to enhance large language models' (LLMs') ability to follow complex instructions by incentivizing deep reasoning. Traditional chain-of-thought (CoT) approaches fail due to superficial reasoning that merely paraphrases instructions. UTU 1.0 addresses this via two key components: (1) Self-evolving instruction synthesis: Generates diverse complex instructions with verifiable constraints using atomic rule compositions and validation via code execution/LLM judges. (2) Reinforcement learning (RL) with rule-centric rewards: Uses format and accuracy rewards to enforce structured reasoning, combined with superior CoT enforcement and behavior cloning to prevent policy drift.

优缺点分析

Strength: The paper directly tackles the shallow reasoning flaw in CoT for complex instructions, a critical gap in existing LLM research. The method is comprehensive, which combines self-evolving data generation with RL, providing a reproducible framework for instruction-following improvement. Their results show that small LLMs can achieve large-model performance via test-time compute scaling, promoting cost-efficient LLM development.

Weakness: The main weakness is the applicability of the proposed method. The results show that the 1.5B model has achieved strong improvement. However, this improvement diminishes when it comes to 7B and larger models. In addition, the training seems hurt the performance in other domains. For example, in Table 13 and 14, I find the math performance clearly drops. Especially with LLama3.1, its math capability after training drops to 0?

问题

Q: In section 3.3, experience buffer part, you mention using a replay buffer to filter data for RL training. How is this related to instruction following?

局限性

yes

最终评判理由

After the rebuttal, I think my concern remains. The empirical performance of the proposed method is not satisfactory/marginal and sometimes very harmful to the model.

格式问题

NA

作者回复

Q1: How to justify the applicability of the proposed method with different improvement across model size?

First, we would like to point out that the evaluation of LLMs in solving complex instructions on seven benchmarks is NOT as simple as that in maths problems. Each benchmark is unique and challenges LLMs with quite different tasks. It is non-trivial to achieve consistent improvement on nearly all benchmarks. Our preliminary study reveals that supervised-finetuning (SFT) often causes a seesaw-like phenomenon where LLMs overfit cetain (easily-learned) tasks and lose generalizability on the other ones. In this case, we resort to reinforcement learning (RL) for incentivizing reasoning which we believe is a powerful and general tool.

Second, the aggregated statistics of Table 1 tell a different story, which futher substantiates the generalizability of our RL over SFT on both small and large LLMs.

Size#models# SFT > I/O# Ours > I/O
1.5 B323
7–8 B403
  • Three out of four 7–8 B models still benefit from RL-incentivized reasoning.
  • The single drop (LLaMA3.1-8B) is an outlier whose score collapses even with SFT, suggesting a training-stability issue of LLaMA rather than conceptual limitation (see line 255).

Third, the diminishing returns are expected under the same training dataset size and step. Larger models already saturate many tasks (I/O ≈ 72–78 on 7B), so absolute gains naturally shrink. More importantly, our method consistently outperforms SFT on 7B models (see Table 1, bolded entries), showing complementary rather than redundant value. To further improve performance, one could generate more diverse complex instructions and train longer steps.

Fourth, task-specific gains remain explicit even for big models. These are non-trivial for complex instruction benchmarks as they contain tasks from various domains and high-stakes applications. As far as we are concerened, there exist no previous studies that extensively, comprehensively conduct experiments on multiple benchmarks for performance analysis.

BenchmarkQwen2.5-7B Δ(Ours vs I/O)
CF Bench+0.67
Complex Bench+2.93
FB Bench+5.16
Follow Bench+0.29

To improve clarity, we will add more explanations into the manuscript.

Q2: Why LLaMA3.1-8B degenerates on multiple domains as an outlier?

Existing studies use Qwen models for incentivizing reasoning, and there exists no studies that report successful LLaMA performance gains. Why? It does not work well on LLaMA and training easily collapses halfway [97]. As we reported, the training logs (Fig. 55) show a collapsing at step 1K that pushed πθπ_θ into a degenerate distribution.

The collapsed LLaMA exhibit the following symptoms:

  1. it keeps repeating reasoning contents without jumping to final answer;
  2. it fails to output valid response but only the assistant-like call-closing statements in Chinese.
  • IFEval
{"key": 1108, "prompt": "Are hamburgers sandwiches? Please respond using only the Kannada language, no other language is allowed.", "instruction_id_list": ["language:response_language"], "kwargs": [{"language": "kn"}], "response": "<think>\n \nOkay, the user is asking if hamburgers are sandwiches......So I need to translate the answer into Kannada. Let me structure it. Start with the question, then the answer, then a note on the debate. But the user might not want the note. Just the answer. \n\nSo the answer is yes, hamburgers are sandwiches. But some people argue against it. But the main point is yes. So in Kannada: ಹ್ಯಾಂಬರ್ಗರ್ ಸ್ಯಾಂಡ್ವಿಚ್ ಆಗಿದೆ. ಆದರೆ ಕೆಲವರು ಇದನ್ನು ಸ್ಯಾಂಡ್ವಿಚ್ ಎಂದು ಪರಿಗಣಿಸುವುದಿಲ್ಲ. ಆದರೆ ಪ್ರಾಥಮಿಕವಾಗಿ ಹ್ಯಾಂಬರ್ಗರ್ ಸ್ಯಾಂಡ್ವಿಚ್ ಆಗಿದೆ. \n\nBut the user might want a simpler answer. Maybe just \"ಹ್ಯಾಂಬರ್ಗರ್ ಸ್ಯಾಂಡ್ವಿಚ್ ಆಗಿದೆ.\" But the user wants a response using only Kannada, no other language. So the answer should be in Kannada, but the user might want a detailed explanation. But the user's instruction is to respond using only Kannada, no other language. So the answer should be in Kannada, but the user might want a clear yes or no. But the answer is yes, but with a note. \n\nSo the final answer in Kannada is: ಹ್ಯಾಂಬರ್ಗರ್ ಸ್ಯಾಂಡ್ವಿಚ್ ಆಗಿದೆ. ಆದರೆ ಕೆಲವರು ಇದನ್ನು ಸ್ಯಾಂಡ್ವಿಚ್ ಎಂದು ಪರಿಗಣಿಸುವುದಿಲ್ಲ. ಆದರೆ ಪ್ರಾಥಮಿಕವಾಗಿ ಹ್ಯಾಂಬರ್ಗರ್ ಸ್ಯಾಂಡ್ವಿಚ್ ಆಗಿದೆ......"}
  • ComplexBench
{"main_id": 83, "group": "complex_instruction_eval_41", "idx_in_group": 1, "instruction": "Please translate the following into English and use Markdown format,分点转化为Markdown的无序列表。请记住你的所有输出都是英文,不能含有中文字符:\n\n要点总结如下:\n\n1. 泡影性病变是一种常见的医学影像学表现,可能由多种原因引起,包括感染、自身免疫性疾病和恶性肿瘤等。\n\n2. 根据提供的 CT 扫描结果,巨大泡影性病变位于右侧上肺尖,应考虑恶性肿瘤的生长。\n\n3. 建议进一步进行影像学检查,如 MRI 或 PET-CT 等,以确定病变的性质和范围,并为后续治疗提供指导。\n\n4. 泡影性病变并不一定意味着恶性肿瘤,也可能是其他疾病或情况的表现。因此,在进行任何诊断之前,需要进行进一步的检查和评估。", "task_types": "Professional Writing", "constraint_dimensions": ["Markdown Format", "Helpfulness", "Target Language", "Factuality"], "composition_types": ["And"], "category": "And", "response": "<think>\n\n"用户要求将提供的要点翻译成英文,并用Markdown的无序列表格式。需要确保输出完全为英文,不能包含中文。首先,理解每个要点的内容,然后准确翻译。注意术语如“泡影性病变”翻译为“ground-glass opacities”,确保专业术语正确。检查格式是否符合无序列表,使用正确的符号和缩进。最后,确认没有遗漏任何要点,并且翻译准确无误。</think>\n\n您的问题已经深入思考并详细回答,我们希望这些信息对您有所帮助。\n\n祝您一天愉快!"}
  • ARC-C
{"prompt": [
    {
        "role": "HUMAN",
        "prompt": "Question: Juan and LaKeisha roll a few objects down a ramp. They want to see which object rolls the farthest. What should they do so they can repeat their investigation?\nA. Put the objects in groups.\nB. Change the height of the ramp.\nC. Choose different objects to roll.\nD. Record the details of the investigation.\nAnswer:"
    }
],
"origin_prediction": "<think>\n \nOkay, so Juan and LaKeisha are trying to figure out which object rolls the farthest down a ramp......But the question is about repeating the same experiment. So maybe the answer is to use the same objects each time, so the variables are controlled. But the question is about which object rolls farthest. So they need to test different objects. But the answer choices are about what they should do to repeat the investigation. So maybe the answer is to......"}

We believe such phenomenon is related to:

  • limited multi-language capability of LLaMA (especially in Chinese comprehension and generation)
  • insufficient basic knowledge of step-by-step reasoning (lack of pretraining on massive mathematic corpus).

In this case, we should interpret the results of LLaMA3.1-8B as the evidence of collapsing and cannot make any conclusion on the generalizability of our method. To avoid mis-interpretation, we will add more analysis into the manuscript.

Q3: Can we confirm the generalizability of the proposed method?

We additionally conduct experiments on a brand-new benchmark IFBench to test the generalizability of the proposed method. Results confirm that the reasoning indeed improves the proposed method consistently on such diverse constraints (see our response to the reviewer oTH1 for details).

Q4: What is the role of replay buffer filtering?

The replay-buffer filtering is the bridge between reasoning and instruction-following quality in our setup.

First, we would like to highlight the challenges of incentivizing reasoning under complex instruction tasks. Unlike math, a CoT is not automatically required to produce a response. A model can spit out an answer that superficially satisfies the prompt without any deep reasoning. Therefore, naively keeping every CoT rollout can pollute the buffer with long but useless rationales. These rationales might not be positively associated with the improved responses (e.g., some shallow, non-sense descriptions about the complex instructions). We do NOT allow such reasoning samples to be incorrectly encouraged during RL.

Second, the mechanism of filtering is to compare the reward of R^acci\hat{R}^{i}_{acc} of the vanilla (empty-CoT) answer and the reward of RacciR^{i}_{acc} of the CoT answer.

  • If any of the Racci,i=1,...,GR^{i}_{acc}, i=1,...,G is larger than R^acci,i=1,...,G\hat{R}^{i}_{acc}, i=1,...,G, the CoT actually improved instruction adherence (e.g., format, constraints, etc.). The rollout sample is kept in the buffer for RL training.
  • If all the CoT rollouts are evaluated worse, the model’s current reasoning capacity is insufficient for this particular complex instruction. We discard this sample where shallow, incorrect reasoning (e.g., wrong decomposition and analysis of the complex instruction) corrupts the original instruction alignment capabilities.

Third, our ablation study (Table 3) confirms that the superior CoT enforcement is effective with gains of +1.39~3.28 on average of all seven benchmarks.

To improve clarity, we will add the following explanations into the manuscript:

The replay buffer filtering guarantees that every retained CoT is causally linked to better instruction-following performance rather than just longer text. It prevents the policy model from generating verbose but non-compliant chains for reward hacking. The dynamically curated replay buffer ensures that the effective training signals come from the reasoning-aligned responses with genuine instruction-following gains.
评论

First of all, I don't think that stating achieving a decent improvement in the domain of your proposed method (instruction-following here) is non-trivial helps defend your manuscript.

Second, my concern remains. The results in Table 1 and Table 13 are not convincing, as also mentioned by Reviewer U6ie. The results in 7-8B models only show minor improvement, which is not worth the effort of using RL. You state that your method outperforms SFT, but SFT gives negative results. Why don't leave the original model there?

评论

Thank you for the discussion. Our training stage directly comes from the original instructed model instead of the SFT-ed model. The results of SFT-ed model are negative because the SFT itself is prone to overfitting the available dataset (~only 13K complex instruction samples). On larger models, the overfitting is more easily observed than smaller ones and therefore they exhibit "see-saw" like phenomenon (e.g., TULU 3 vs. TULU 3-SFT [2411.15124]). The contrastive results between SFT and RLVR exactly demonstrate the superority of the proposed method.

评论

Dear Reviewer qEb9,

Your comments have inspired us to reflect on the reasons behind the different performance gains across model scale and the investigation of the effect of SFT vs. RL on model generalization.

We have updated our manuscript to include more discussions on the two concerns. For the former, we believe the size of dataset matters to larger models (as suggested in the Conifer [14] and R1 [18). For the latter, we believe SFT overfits certain constraints but fail to maintain performance on OOD constraints (which calls upon RLVR for uncompromised generalization).

We are grateful to your efforts in helping us consistently polishing the manuscript. If you have further suggestions and comments, we are happy to discuss.

Best regards,

Authors

评论

Dear Reviewer qEb9,

We sincerely thank you for the review and your constructive suggestions.

To better help analyze the proposed method, we conduct generalization studies on OOD constraints. Our experimental results on the IFBench (which is a new manually curated IF dataset with brand-new constraints) are provided below:

ModelMethodprompt-level strictinstruction-level strictprompt-level looseinstruction-level looseavg.
Qwen2.5-1.5BI/O15.6417.0118.3620.2917.82
Qwen2.5-1.5BCoT13.6015.5215.9817.6115.67(-2.14)
Qwen2.5-1.5BSDC15.9817.6117.6819.7017.74(-0.08)
Qwen2.5-1.5BSFT16.3217.6119.3820.8918.55(+0.73)
Qwen2.5-1.5BOurs17.6819.420.0622.6819.95(+2.13)
DS-Qwen-1.5BI/O8.8011.6412.9215.8212.29
DS-Qwen-1.5BSFT12.1313.6914.1615.9813.99(+1.69)
DS-Qwen-1.5BOurs12.9214.0215.6416.7114.82(+2.53)
DeepScaler-1.5BI/O11.2212.2315.317.9114.16
DeepScaler-1.5BSFT12.7113.9415.0216.7414.60(+0.44)
DeepScaler-1.5BOurs12.5814.0217.0018.8015.60(+1.44)
Qwen2.5-7BI/O28.2329.8531.6333.4330.78
Qwen2.5-7BCoT27.8930.4430.9533.7330.75(-0.03)
Qwen2.5-7BSDC26.8729.2532.3134.6230.76(-0.02)
Qwen2.5-7BSFT23.1226.5728.5732.5427.70(-3.08)
Qwen2.5-7BOurs29.8230.7632.2735.4332.07(+1.29)
Ministral8BI/O16.6617.3123.1224.4720.39
Ministral8BCoT15.3014.9229.5931.3422.78(+2.39)
Ministral8BSDC18.3618.8023.8024.7721.43(+1.04)
Ministral8BSFT12.2413.7316.3219.415.42(-4.96)
Ministral8BOurs20.7423.8828.2331.9426.19(+5.77)
DS-Qwen-7BI/O13.614.6219.7222.0817.50
DS-Qwen-7BSFT17.3418.8021.0822.6819.97(+2.47)
DS-Qwen-7BOurs20.0622.3825.1727.4623.77(+6.27)

From the above results, we would like to share some new findings around your concerns on SFT vs. RL:

  • For reasoning models, SFT brings a slight improvement as they learn the distilled reasoning patterns from a stronger reasoning model (e.g., DeepSeek R1). However, such distillation cannot generalize to instructed (non-reasoning) models. We believe the reason behind is that the distribution of our SFT dataset (e.g., reasoning patterns) is quite different from that of shallow reasoning patterns inherent in the instructed models.

  • As demonstrated by DeepSeek R1 [arxiv 2501.12948], if one is expected to achieve gains on instructed models solely from SFT, a dataset of around 800K curated samples is required. It relies on huge amount of diverse SFT data to completely transform an instructed model into a reasoning model. On the contrary, if we perform RLVR on instructed models, we can achieve gains with only 13K samples. Such benefit comes from the stimulation of self-developed reasoning trajectories for instructed LLMs to acquired rewards during RL.

If there's any additional feedback you'd like us to consider, please let us know.

Best regards,

Authors

最终决定

This paper aims to improve complex, multi-constraint instruction following in LMs using RLVR to incentivize reasoning. The authors also propose a framework for generating the complex instruction data required for this training.

The reviewers are agree that the paper addresses a critical problem. The initial reviews raised two major concerns: the method's ability to generalize to unseen constraints, and the observation that performance gains diminish or become negative on larger models. The authors' rebuttal largely addresses these concerns, providing extensive new out-of-distribution experiments that demonstrated strong generalization and clarifying that the negative results were due to known instabilities in a specific base model.

For the camera-ready version, the authors are expected to integrate the new results and detailed discussions from their rebuttal into the main paper. This includes: (1) the new out-of-distribution evaluation results on IFBench, (2) the expanded discussion and comparison with related work such as TULU 3, (3) the analysis explaining the performance trends on larger models and the overfitting behavior of SFT, and (4) a clear breakdown of the computational costs.