PaperHub
6.4
/10
Poster4 位审稿人
最低3最高5标准差0.7
4
4
3
5
4.0
置信度
创新性3.0
质量2.5
清晰度2.8
重要性2.3
NeurIPS 2025

Steering Information Utility in Key-Value Memory for Language Model Post-Training

OpenReviewPDF
提交: 2025-05-07更新: 2025-10-29
TL;DR

We introduce a method that encourages LMs to leverage their pretrained knowledge during post-training.

摘要

关键词
InterpretabilityInterventionModel SteeringLanguage ModelsPost-Training

评审与讨论

审稿意见
4

This paper introduces InfoSteer, a lightweight method that improves post-training of language models by steering toward underutilized FFN memory vectors. It enhances performance and interpretability across models and tasks.

优缺点分析

Strengths

  1. The method is novel, lightweight and easily integrable into existing pipelines.

  2. Extensive experiments across multiple model families and tasks demonstrate improvements

Weaknesses

  1. Motivation is underdeveloped: It's not entirely clear why encouraging a more uniform (higher-entropy) key distribution is beneficial. In practice, a sharp key distribution might indicate the model is confident in retrieving task-relevant information, while low-activation keys may represent irrelevant memory. Forcing these less relevant keys to contribute might dilute useful signals. In the extreme case, a uniform distribution means every token receives the same memory mix, which seems counterproductive. A deeper justification is needed here.

  2. Information Flux definition (Section 6.2): Using token-level entropy over key coefficients as a measure of “Information Flux” seems to be unconventional. It’s unclear whether this term has any precedent in prior work. The paper should explain why this metric reflects self-interpretability. And a comparison with the baseline SFT model on this interpretation is needed. Also, why do operators exhibit lower flux—could this be due to syntactic predictability rather than meaningful information usage?

问题

  1. As in Weakness 1, why is a more uniform key distribution necessarily desirable? Could this harm precision by encouraging irrelevant information use?

  2. If the main goal is to increase key entropy, what happens if we apply the intervention-only strategy during inference? Would that still yield benefits?

局限性

N/A

最终评判理由

The rebuttal addressed some of my concerns, and after reading other reviewers' comments, I believe that 4 is a reasonable score.

格式问题

The authors include a Limitations section, which is commendable. However, the discussion is relatively brief and focuses more on future directions than on critical analysis of assumptions or risks. For instance, the core assumption that increasing key entropy improves information use lacks justification. Similarly, the proposed Information Flux metric is not clearly grounded.

作者回复

Thank you for your constructive review of our manuscript with a positive assessment of our work. We noted your attention to a finer-grained discussion about the key-distribution in FFNs and clarification re IF score. We will your concerns as follows:

About Motivation for Higher-Entropy Key Distribution

Thank you for the great question! The reviewer raises a question about why encouraging more uniform key distributions is beneficial. We clarify several key points:

  • Not Uniform Distribution, But Less Extreme Low-Entropy Distribution: We first want to clarify that our motivation for promoting a higher-entropy key distribution is not to achieve a perfectly uniform distribution, as that would indeed be another extreme. Instead, our objective is to introduce a "damp factor" to the extreme low entropy distribution introduced in vanilla SFT. We think the extreme low entropy is harmful and leads to detrimental underutilization of parametric knowledge (only a very small portion of memory vectors are used). In practice, we find after training with infosteer the entropy is close to that of the pre-trained models, while vanilla SFT would have a very low key distribution (H(pretrained LM) ~= H(Info-steered SFT) >> H(vanilla SFT)).
  • Learning with Intervention, Not Intervention Alone: Following up the clarification, our main goal is to not leverage the steering-only method to work (re Q2), instead, the steering specifically designed to help or assist post-training methods like SFT (SFT -> Info-steered SFT). We know that the issue of SFT is that it will lead models to mimic the pattern of a small curated post-training datasets so easily overfit. With the help of the steering method, it feels like adding a weighted damp factor to prevent the model blindly fit this small curated dataset pattern but learn in a more “global” loss landscape.

[Empirical Evidence] A straightforward experiment is that vanilla SFT performs worse on OOD tasks, but info-steered SFT generalizes better to OOD situations. The results across three model families on 15 downstream tasks prove the effectiveness of our methods.

We will dedicate a discussion section in the limitation section and main body in the camera-ready version to ensure this is clarified for readers.

About Information Flux Definition and Comparison with Baseline

Thanks for asking this! The IF part is quite interesting, and we would love to discuss it further:

  • Metric Definition and Precedent. You are correct that Information Flux is our proposed metric. We define it as lH(k^(l))\sum_{l}{H(\hat{k}^{(l)})} where k^(l)\hat{k}^{(l)} is the normalized key coefficient distribution at layer ll. This measures the diversity of memory vectors engaged during token generation—higher IF score indicates more memory vectors are engaged during the process. This name may be new, but using cumulative entropy across layers is indeed having some precedent literature[2][3].
  • Why This Reflects Self-Interpretability. The IF score reveals which tokens require diverse vs. narrowed knowledge access. The case you mention in W2 (syntactic predictability) is precisely the point—for those syntactic tokens, models do not need many knowledge sources (low IF score), so they easily make the prediction. For some hard tokens, the model needs to "think harder" to use more memory vectors to make decisions. We think the IF score reveals how LMs deal with different types of token generation. Though it seems not quite comparable to mech interp methods like SAE/circuits/neurons, we think it provides a different angle to understand what happens in the model itself.

[Significance of IF Score Before/After SFT]

This is a very good question to ask! We indeed have these scores and report the results when noticing a difference. Here are 1K samples from Ultrafeedback and we highlight the significance below.

Token TypeBase ModelVanilla SFTInfo-steered SFTDifference
Content words2.03 ± 0.152.05 ± 0.162.67 ± 0.21+0.62
Function words2.01 ± 0.142.02 ± 0.151.73 ± 0.12-0.29
Punctuation1.98 ± 0.132.00 ± 0.141.31 ± 0.09-0.69

We believe this represents a clear difference before/after Info-steered SFT, so we report this interesting case study in our main paper. We will add this table to the camera-ready version to quantitatively highlight the difference. Thank you once again for your valuable feedback!

Reference

  • [1] Kim, Jiyeon, et al. "Knowledge Entropy Decay during Language Model Pretraining Hinders New Knowledge Acquisition." The Thirteenth International Conference on Learning Representations.
  • [2] Ali, Riccardo, et al. "Entropy-lens: The information signature of transformer computations." arXiv preprint arXiv:2502.16570 (2025).
  • [3] Chang, Hoyeon, et al. "How do large language models acquire factual knowledge during pretraining?." Proceedings of the 38th International Conference on Neural Information Processing Systems. 2024.
审稿意见
4

This work posits that post training (SFT) currently makes subpar use of knowledge acquired during pre-training and suggest two steering methods to improve upon that, including

  1. intervening on memory vectors with low corresponding key values during the forward pass and
  2. regularizing the entropy of keys in the gradient flow during backpropagation, both which help to encourage more balanced activation of memory vectors ( as opposed to overfitting on a few ).

Empirically the authors show the two methods give generally improved performance against SFT on various benchmarks/tasks across 3 models (of various sizes 1B to 8B parameters) including for in and out of distribution settings.

The paper provides ablations on hyper parameters introduced by the methods, and analysis on distribution shift of key values observed post steering. The authors additionally propose a token attribution interpretability method based on token-level entropy over the key coefficient distributions as a proxy for how many memory vectors are activated during token generation and provide some qualitative examples of token level highlighting induced by the method for interpretability purposes.

优缺点分析

Quality: The overall quality is strong though there are a few areas that need improvement:

  1. In the experiments while its shown the methods do better than vanilla SFT, the proposed methods are not compared against any other SFT (Lora, ReFT, etc) or steering baselines such as inference time interventions (ITI) or mean difference vector steering ( Arditi 24 et al ).
  2. The arithmetic ID/OOD experiment of using the arithmetic GSM8K to post-train and then doing In Distribution results for it and then Out of Distribution results for 5 other arithmetic tasks seems like a misnomer to me.
    They collectively are different tasks, but still within the domain of “arithmetic”. By this sort of logic the authors could have re-used the model checkpoints from table 1 to calculate “out of distribution” results for each of the 8 tasks relative to the other ones ( ie, you’ve done vanilla SFT and steering w/intervention, w/regularization on BoolQ so you could see how those models each do on average against the 7 other tasks (PIQA to OBQA).
  3. In 5.2, the claims in “Insufficient Utilization of Pretrained Knowledge in Modern LLMs” seem a little over stated as written. Saying SFT methods “substantially underutilize knowledge” because Gemma-2-9B base model achieves 90.1% on HellaSwag while reaches 95.7% with intervention steering is comparing against the base model and not the actual baseline vanilla SFT (which gets 94.8% which is within a point). Without CIs its difficult to tell (i’m assuming the green and blue colors reflect this somehow), but for Gemma-2-9B it looks like of the 8 tasks, only two show some improvement?
  4. No mention or explanation is given for why the methods do worse both for OBQA for Qwen1.5B & SIQA for LLama 1.B?
  5. Discussion of where and how the methods improved performance and why would be appreciated

Clarity: The motivation paper is well motivated and mostly well written and easy to follow, but some sections could benefit for clarity:

  1. Section 5.4
  • I didn’t see where it was mentioned what datasets are used for each group task (Reading Comp, Knowledge, etc ) and
  • the average delta’s seem off. For instance, for Commonsense Reasoning, steering improves 2.4 and 1.8 respectively over Vanilla SFT (which implies 2.1) but the Average Delta says its 2.3 ??
  1. Section 6.1
  • What do low key, medium key and high key regions mean exactly here? ie, what are the cutoffs for considering this? This section is interesting, but needs to include more on the setup and explanations. For which datasets are these shown?
  1. Section 6.2: The section discusses how “LMs might be more self-interpretable via proper guidance without extra training” , but then does instruction tuning ( and its unclear if this is with vanilla SFT or using which of the two interventions ) and then does a post hoc interpretability method involving token entropy, which seems to me to fall into the camp of post-hoc interpretable feature importance methods (Integrated Gradients for instance) which do token highlighting based on model gradients without extra training. I think the proposed Information Flux method is interesting and hope it gets flushed out more systematically/rigorously in a separate work ( as it stands here, with the few qualitative examples provided its hard to see any clear advantage of it or quantitative trend )

Significance: The results of steering with either of the methods generally seems to improve, but without more study comparing against other baselines for knowledge intensive tasks its hard to tell.

Originality: Using steering methods to promote key/memory vector utilization explicitly seems novel to me from what I’ve seen in the literature, and the interpretability method proposed in 6.2 also seems novel .

问题

Please see questions raised in the strength and weaknesses section.

Adding comparisons against some additional baselines, clarity improvements and going more into analysis of the results seen ( when are they successful and when not ) would help strengthen the paper.

Additionally, did the authors try combining the two methods together ( which would make logical sense) and if so what results did you see?

局限性

yes

最终评判理由

The additional experiments and explanations provided help addressed most of my concerns. I do think comparing the combination of the two methods together is the biggest open point for me since its still unclear when to use one method vs the other ( but regardless you've convincingly shown they do improve upon SFT alone and other methods ( though I think ITI and Diff MEANs could also be used on top of SFT, but with additional data needs ) . With that I update my rating from 3 to 4.

格式问题

none

作者回复

We thank you for your detailed and constructive review of our manuscript! We are also grateful for your specific suggestions, which we believe will significantly strengthen the paper. Specifically, we want to address key feedbacks below:

[Key Concern] About Comparison with Additional Baseline:

We first kindly clarify that InfoSteer is not a replacement for post-training methods like LoRA or ReFT, nor is it a test-time algorithm like ITI or mean difference vector steering. Instead, InfoSteer is a steering technique that can be applied to enhance post-training methods (SFT -> info-steered SFT).

That said, we acknowledge the value of empirical comparisons and have included additional baselines* (LoRA[1], ReFT[2], ITI[3] and Diffmean[4][5]) for 7B scale model in our results:

ModelTraining MethodBoolQPIQASIQAHellaSWinoGARC-eARC-cOBQA
Qwen-2.5-7Bbase model68.981.277.687.580.379.865.177.6
+ LoRA*71.884.280.991.883.783.868.982.3
+ ReFT*72.184.581.292.184.084.069.282.5
+ ITI*70.583.179.890.982.882.967.881.4
+ DiffMean*70.282.879.590.682.582.667.581.1
+ vanilla SFT72.484.981.592.484.284.269.682.8
+ steered SFT w. intervention74.186.381.895.187.286.273.784.2
+ steered SFT w. regularization76.485.781.094.386.585.472.983.4
LLaMA-3-8Bbase model70.385.675.790.881.986.275.380.5
+ LoRA*74.188.779.393.185.189.979.885.2
+ ReFT*74.389.079.693.385.390.280.185.5
+ ITI*72.887.478.292.084.288.778.584.1
+ DiffMean*72.587.177.991.883.988.478.283.8
+ vanilla SFT74.689.379.993.585.690.580.485.8
+ steered SFT w. intervention77.190.282.096.387.492.481.687.5
+ steered SFT w. regularization76.589.581.295.686.891.780.986.8
Gemma-2-9Bbase model71.686.377.290.182.587.577.881.7
+ LoRA*73.889.581.294.386.491.281.586.0
+ ReFT*74.089.881.494.686.691.481.786.2
+ ITI*72.988.780.393.585.790.380.885.1
+ DiffMean*72.688.480.093.285.490.080.584.8
+ vanilla SFT74.390.181.794.886.991.782.086.4
+ steered SFT w. intervention77.291.883.195.788.593.583.488.2
+ steered SFT w. regularization76.590.982.494.987.892.882.787.5

LoRA and ReFT perform similarly to vanilla SFT, with slight variations depending on the task. ITI and DiffMean (inference-time methods) generally perform worse than training-based approaches. Info-steered SFT methods consistently outperform all baselines, demonstrating the value of information steering during post-training. We will incorporate these baseline into the paper to strengthen its position.

About Arithmetic ID/OOD Experimental Design:

Thank you for mentioning this! We agree this terminology could be clearer. We think if we look at those data even if they are mostly arithmetic, the question narrative and domain may be quite different. But we also agree with reviewers that, esp in the LLM era, the term “distribution” should be in a stricter inspection.

To address the reviewer's concern more directly, we implemented an even stricter evaluation strategy than suggested. We took Llama-3-8B checkpoints specifically trained on OBQA and evaluated them on all five arithmetic datasets to test cross-domain transfer:

ModelID Eval (OBQA)OOD Eval (Arithmetic)
Base Model80.568.2
+ Vanilla SFT85.8 (+5.3)61.4 (-6.8)
+ Steered SFT w/ interv.87.5 (+7.0)68.4 (+0.2)
+ Steered SFT w/ reg.86.8 (+6.3)68.1 (-0.1)

This stricter evaluation reveals that vanilla SFT dramatically hurts cross-domain generalization (-6.8%), while adding our steering methods maintain the overall generalization performance. We will add these results into the main body, thank you for your suggestions!

About Interpretation of the Performance

Out of 96 experiments (8 benchmarks × 3 model families × 2 model sizes × 2 intervention methods), we observe consistent gains in 94 results, suggesting that language models may not fully realize their potential from pre-training alone. However, we agree that we should adopt a more humble and scientific tone in our paper. We will revise it to use modest language that objectively reports the gains our method achieves.

Re the failure cases you mentioned, we examined those two benchmarks and found that the entropy of key distributions before steering in FFNs was abnormally higher (around 2.97) for these models compared to others (around 2.03). We believe this is an interesting phenomenon that may be worth investigating further—specifically, whether considering prior key entropy could enable more fine-grained steering. We will add this discussion to the main body of the paper. Thank you for raising this excellent question!

About the Potential Combination of the Method

Yes, we have tried to add two pytorch hooks and work together, in some preliminary experiments we find it showed marginal improvements over the individual method so we do not explore that further.

About Details in Experiment

We summarize all comments related to details about our experiment here.

[Key Distribution Shift] We report the average results across all 15 datasets previously mentioned in this analysis section. The low/medium/high key regions are defined by percentile-based cutoffs: 0-25th, 25th-75th, and 75th-100th percentiles, respectively. Due to page limits, we do not disclose all information in the main body. We will add a detailed description and analysis for this section in the appendix.

[Dataset Picked in Different Domain] Based on the dataset characteristics, we categorize them as follows: Reading Comprehension includes BoolQ (yes/no questions based on passages); Knowledge encompasses ARC-e and ARC-c (science knowledge questions) and OBQA (factual knowledge questions with reference material); Commonsense Reasoning covers PIQA (physical interaction reasoning), SIQA (social interaction reasoning), and HellaSwag (sentence completion requiring commonsense); Math includes all arithmetic datasets: AddSub, MAWPS, MultiArith, SingleEQ, SVAMP, and GSM8K (various forms of arithmetic and math word problems); and Linguistic uses the linguistic portion of WinoGrande (pronoun resolution requiring syntactic understanding). When averaging the score there are some cumulated rounding errors (as you noticed) which leads to confusion of this, we will report the larger precision score to avoid the confusion.

We will add these details into our manuscript to make it clarified for readers.

About Section 6.2 Interpretability Claims

We are glad you find this section interesting (as do we!). You are correct that we should at least dedicate a section to give them more attention, but due to limited page space we can only squeeze this one in. A quick response to your question is: both methods see the phenomenon, but the one using entropy regularization shows the most significant results, and we report these qualitative results in the paper. We also have rigorous quantitative experiments to highlight their statistical significance, would love to include that into the camera-ready for better readability.

We sincerely appreciate your suggestions and will incorporate them to enhance our work further. Thank you once again for your valuable feedback!

Reference

  • [1] Hu, Edward J., et al. "LoRA: Low-Rank Adaptation of Large Language Models." International Conference on Learning Representations.
  • [2] Wu, Zhengxuan, et al. "ReFT: representation finetuning for language models." Proceedings of the 38th International Conference on Neural Information Processing Systems. 2024.
  • [3] Li, Kenneth, et al. "Inference-time intervention: eliciting truthful answers from a language model." Proceedings of the 37th International Conference on Neural Information Processing Systems. 2023.
  • [4] Turner, Alexander Matt, et al. "Steering Language Models with Activation Engineering."
  • [5] Arditi, Andy, et al. "Refusal in language models is mediated by a single direction." Proceedings of the 38th International Conference on Neural Information Processing Systems. 2024.
评论

Thank you for your additional experiments and explanations.

I do think comparing the combination of the two methods together is the biggest open point for me since its still unclear when to use one method vs the other ( but regardless you've convincingly shown they do improve upon SFT alone and other methods ( though I think ITI and Diff MEANs could also be used on top of SFT, but with additional data needs ) . With that I'll update my rating from 3 to 4.

评论

Yeah that's a very interesting take if we could combine the two methods to boost the performance further. Also, it would be extremely interesting to see if all those test-time algorithms like ITI could be used to boost post-training (tho we do need to create contrastive pairs as you mentioned). That would be quite beneficial to the whole community.

Thank you again for your constructive feedback and sincere discussion with us!

审稿意见
3

This paper introduces InfoSteer, a method designed to enhance the utilization of pre-trained knowledge in language models (LM) during the post-training phase. InfoSteer treats the Transformer's feed-forward network (FFN) layer as an associative key-value memory and employs two strategies to promote the active use of stored memory vectors. Experimental results demonstrate that InfoSteer improves performance across multiple model families and over 15 downstream tasks. It also shows that the steered models can adaptively allocate information resources, focusing more on generating semantically meaningful tokens while using fewer resources on simple transition tokens.

优缺点分析

Strength

  1. Innovative model steering approach. InfoSteer offers a novel method for post-training by interpreting the FFN layer as a key-value memory structure.

  2. Comprehensive experimental validation: The authors conduct extensive experiments across multiple model series and a wide range of downstream tasks.

  3. Insights into model behavior: Beyond performance improvements, the paper provides valuable insights into how the model allocates resources during token generation. It shows that InfoSteer encourages the model to focus computational resources on semantically rich tokens, offering a deeper understanding of the model's internal behavior.

Weakness

  1. The study lacks comparison to widely-used approaches like reinforcement learning fine-tuning (RLHF). Vanilla supervised fine-tuning (SFT) alone cannot represent the state-of-the-art model capabilities.

  2. As shown in Table 1, the differences between various SFT methods diminish as the model size increases. I'm particularly interested in the performance of the proposed method on models around the 30B parameter scale. Or the authors need to offer some explanations for this phenomenon.

问题

See weakness.

局限性

N/A

最终评判理由

I have re-examined this paper and will stand by my previous comments.

格式问题

N/A

作者回复

We appreciate your thoughtful and constructive review of our manuscript. We’ve noted your main concerns regarding this paper mainly focusing on SFT instead of RL and correlation between performance gain and model size. We try to address your concerns as follows:

About Comparison to RL

We want to kindly clarify that infosteer is not a post-training method compared to RL, but a plug-and-play method to steer both RL and SFT those post-training methods. SFT can be seamlessly incorporated as information-oriented SFT (SFT -> info-steered SFT), which is the same as RL (RL -> info-steered RL).

As stated in the limitation section, in this work as an explorer of this steering method in post-training we primarily focus on the SFT. But we also understand that recently RL become a hot topic which people find it is effective on verifiable domain like math, We also add a preliminary experiment to use infosteer on RL with:

MethodAIME 2024/2025AMCMATH-500MinervaOlympiadAverage
BASELINE
Qwen2.5-Math-7B11.5 / 4.931.343.67.415.619.0
SFT METHODS
SFT22.2 / 22.352.882.640.843.744.1
InfoSteer SFT (interv.)23.2 / 23.455.085.842.645.645.9 (+1.8)
InfoSteer SFT (reg.)22.8 / 22.954.285.041.844.845.1 (+1.0)
RL METHODS
RFT (GRPO)25.1 / 15.362.084.439.346.845.5
InfoSteer RFT (interv.)29.2 / 19.467.488.743.250.149.6 (+4.1)
InfoSteer RFT (reg.)28.4 / 18.666.287.542.048.948.8 (+3.3)

The RFT training was conducted using GRPO for 500 steps with 8 rollouts per prompt. The training dataset consisted of approximately 46k diverse math reasoning problems from a length-filtered subset of the OpenR1-Math-220K dataset, with each problem equipped with demonstrations generated by DeepSeek-R1.

InfoSteer consistently enhances both SFT (+1.0 to +1.8 points) and RL methods (+3.3 to +4.1 points) across all mathematical reasoning benchmarks, with InfoSteer RFT achieving the best overall performance of 49.6 average score. While these experiments are preliminary and cannot make a conclusive justification before conducting whole pack of rigorous ablation, we believe info-steered RL is a very interesting and promising direction for our future work.

About Performance Differences with Model Size and Experiment on 30B Models

We appreciate your observation on the different performance with different model sizes. We may not have the resources to reproduce all pack of results on 30B models in this short period, but we are glad to offer some of the discussion on this Q here:

  • Higher Baseline Performance in Larger Models Make Less Room for Improvement: Larger base models generally exhibit significantly higher baseline performance. The "gap" for improvement through post-training techniques naturally narrows. Thus we think a smaller percentage gain on a very high baseline can represent a substantial absolute improvement in capabilities. For instance, a 5.5% gain on a LLaMA-3-8B model with an already high baseline (e.g., 90.8% on HellaSwag for vanilla SFT vs. 96.3% with steered SFT w. intervention) is still a highly significant and impactful improvement, showcasing InfoSteer's ability to unlock latent potential even from highly optimized models.
  • Substantial Gains Remain Consistent: Even with higher baselines, our method continues to deliver meaningful improvements. 48 out of 48 configs (8 benchmarks x 3 8B models x 2 methods) deliver consistent gain. 37 out of 48 configs achieve +1.5% gain. And OOD exp is also done on 8B models, validating our methods’ effectiveness on extended distribution. All results are reported on avg score of three separate runs.

We will elaborate on these enhancements in the revised manuscript to provide readers with a comprehensive understanding of our findings. Thank you once again for your valuable feedback!

评论

Dear Reviewer hS2q,

We appreciate your thoughtful and constructive review of our manuscript. Re your key concerns about infosteer on RL, we have added additional experiments for info-steered RL during this period. Our preliminary results see that info-steered RL achieves even stronger performance compared to info-steered SFT. We have also included additional baselines (see our rebuttal to Reviewer SHNo). We believe your suggestions further strengthens our work's position, and steering RL may be an interesting and valuable direction for future work.

As the rebuttal period draws to a close, we appreciate knowing whether our response adequately addresses your concerns. Your support during this period would mean a lot to us!

审稿意见
5

The paper provides a new perspective on the role of the FFN layer in LMs. It hypothesizes that the post-training underutilizes the knowledge in memory vectors. By encouraging the usage of the memory vectors during post-training, the resulting LM can have better performance on in-domain and out-of-distribution domains.

优缺点分析

Strengths: (1) The writing is clear. It provides sufficient context and motivation for the method. (2) Convincing results and the motivation behind the approaches. The approach is effective across tasks and pre-trained models. It is also simple to implement.

Weakness: Can't think of any.

问题

  1. The approach shows improved performance during post-training. Can the same approach to applied during pre-training to encourage the usage of the memory vectors? Why is it particularly relevant during post-training.
  2. The post-training data size is relatively small. Is it a reason that the info steering approach is effective, because the model cannot learn the optimal key distribution by itself?

局限性

yes

最终评判理由

I accepted the author's rebuttal and would like to retain my rating.

格式问题

No

作者回复

Thank you for your recognition of our work! We are glad that you find our approach easy to implement and effective across tasks and pre-trained models.

You've raised two questions regarding the applicability of our approach during pre-training and its particular relevance to post-training, especially related to the size of the post-training data. We think these two questions are inherently connected and we will address them together.

During the Post-training Phase (Small Data):

In the post-training phase we typically involve much smaller, more curated datasets (yes your take is correct in Q2!) aimed at aligning the model with specific downstream tasks or human preferences. Here, the key concern is overfitting caused by SFT on this small dataset. Standard SFT on limited data can cause the model to aggressively mimic the patterns of this curated dataset, potentially leading to an undesirable shift in its key distribution towards even lower entropy as we discussed in Sec 6.1. This can result in the underutilization of the rich, generalized knowledge acquired during pre-training, harming the model's ability to generalize, especially to OOD scenarios.

During the Pre-training Phase (Large Data):

You also raise an interesting question about whether the steering method could be applied during the pre-training phase. Previous research has observed that the entropy of key distributions also tends to decrease over pre-training steps [1]. Additionally, researchers have also noticed the phenomenon of attention matrix collapse or attention sink (low "key" entropy in attn matrix)[2][3]. Many studies aim to resist this phenomenon to enable better use of attention parameters. We think this research shares similar intuition with ours but focuses more on the attention module. We would be glad to view this direction as an interesting future work. Thank you for the great questions!

Reference

  • [1] Kim, Jiyeon, et al. "Knowledge Entropy Decay during Language Model Pretraining Hinders New Knowledge Acquisition." The Thirteenth International Conference on Learning Representations.
  • [2] Gu, Xiangming, et al. "When Attention Sink Emerges in Language Models: An Empirical View." The Thirteenth International Conference on Learning Representations.
  • [3] Voita, Elena, et al. "Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, the Rest Can Be Pruned." 57th Annual Meeting of the Association for Computational Linguistics. ACL Anthology, 2019.
最终决定

This paper introduces InfoSteer, a lightweight method to enhance language model post-training. The authors claim that standard supervised fine-tuning (SFT) often leads models to underutilize their pre-trained knowledge by overfitting to a narrow set of activations. InfoSteer addresses this by treating FFN layers as key-value memories and applying steering techniques to encourage a more balanced use of this stored information. The method is shown to deliver consistent performance gains across diverse models and tasks.

The paper's primary strengths are that the technique is interesting, simple, and empirically effective, showing consistent improvements in both in-domain and out-of-domain settings. However, the initial submission raised some concerns among reviewers. During this period, the authors had to conduct a substantial amount of new work, including: adding comparisons against four crucial baselines (LoRA, ReFT, ITI, DiffMean), performing a new and stricter out-of-distribution (OOD) experiment to properly validate generalization, and providing quantitative backing for their interpretability metric. The evaluation maintains a controlled comparison by applying the same optimization parameters and single-epoch training across all methods. While this effectively isolates the gains from InfoSteer, it also implicitly assumes this configuration is fair to the baselines, which might perform better with individually tuned hyperparameters.

Furthermore, while the authors provided promising preliminary results applying their method to reinforcement learning, this was only in response to a reviewer's comment Given that RL-based methods are a dominant paradigm for post-training and the authors' own results suggest even stronger gains in that setting, the lack of a complete RL evaluation leaves a key claim of general applicability under-explored.

Overall though, the paper was strengthened throughout the review process and the new results show the method performs favorably compared to other approaches, have better generalizability, and seems to be even more effective in RL.