Parameter-Efficient Fine-Tuning of State Space Models
摘要
评审与讨论
This paper studies the fine-tuning of state-space models, in particular, S4 and S6. Empirical studies on fine-tuning the encoder and the decoder using many different existing tuning mechanisms are shown. Then, an SDT-P fine-tuning strategy of the autoregressive modules is proposed based on pruning and then sparsely fine-tuning a set of latent states. It is also empirically shown that the pruning stage can be pruned. Experiments show that the SDT method works better than the traditional LoRA strategy.
update after rebuttal
I have raised my score to weak accept to reflect my current position.
给作者的问题
- Lemma 1 is about the expressiveness of different ways of doing the fine-tuning. It does not consider, for example, the stability of representations, or more naively, how easy it is to reach the target by only fine-tuning the encoder. Can the author(s) comment on this?
- Lemma 1 seems to diminish the significance of the proposed method, as it says that fine-tuning can be done without changing the autoregressive units. This may be related to the first point, but can the authors provide more clarification on this in the manuscript so that the transition from section 4 to 5 is smoother?
- On line 279, the author(s) wrote "assume all hidden dimensions are active." What does that mean precisely?
- What is the efficiency of the proposed method compared to, for example, LoRA? For example, Figure 2 is mainly a comparison based on the same number of tunable parameters. What about a comparison based on the fine-tuning time? Also, how do these change as the sequence lengths grow?
- How are and in Algorithm 1 selected in Table 3 and 4 and why is it a fair comparison to LoRA with whatever the hyperparameters you selected there? These are necessary to be discussed in the main manuscript.
论据与证据
I find the claims about the empirical advantages of SDT well-supported by experiments.
方法与评估标准
I find the proposed methods and evaluation criteria appropriate for the problem at hand.
理论论述
I did not check the correctness of the proofs. In fact, I am not satisfied with the presentation of some theoretical statements, which I will clarify in the comments/questions sections.
实验设计与分析
I checked the soundness of the experimental designs and found them appropriate.
补充材料
I did not review the supplementary material.
与现有文献的关系
This paper analyzes the fine-tuning of a promising class of large language models, which is certainly an important problem.
遗漏的重要参考文献
NA
其他优缺点
Strengths:
- The problem studied in this paper is important and opportune.
- The empirical study is comprehensive and convincing.
Weaknesses:
- While the central ideas of the paper are clear, I do have some comments on the presentation of the material. See the section below.
- The theoretical statements are on the rough side and many statements need rephrasing. Please see the section below.
其他意见或建议
The comments here are mainly about the presentation of the material and the delivery of the theoretical statements. My current evaluation of "weak reject" is mainly due to the presentation issues. I can guarantee that once the author(s) put efforts into addressing the following comments, pushing up my evaluation would be an easy reach.
-
Right now, the fine-tuning strategies studied and compared in this paper are only referred to without any other details provided. While some of them are popular, readers who are not experts on fine-tuning may find it hard to parse all of them, and consequently Table 1, all at once. It would be useful to add a brief and mathematical introduction to some of them. An introduction in the supplementary material would be otherwise useful.
-
Lemma 2 and Theorem 1 need rephrasing to be mathematically correct. In particular, consider the following questions:
- In Lemma 2, the quantity in (5) is just a number and does not depend on the initial model . This does not seem to correctly reflect the definition of (5).
- In Lemma 2, when saying "the minimum number of tunable parameters," do you mean the precise minimum number (so that we know for sure that if the number goes below it, then we'll not be able to do the job) or an upper bound of the minimum number?
- In Theorem 1, and are not defined. I think they come from the target model. In that case, it needs to be clearly stated what the target model is.
- In Theorem 1, when saying "selectively fine-tuning," what is the training algorithm being used? Or is it just an algorithm-free statement that considers the "best" way of modifying the tunable parameters?
- In Theorem 1, what does "accurately represent" mean?
-
Some minor comments:
- Technically, an S4 model refers to a model where is the sum of a diagonal matrix plus a rank-one matrix, whereas an S4D model is the one with a diagonal . The author(s) should be more precise.
- Throughout the manuscript is used for entrywise product. This is a very misleading notation as it is often used for the Kronecker product. Consider using or for the Hadamard product instead.
- On line 293-294, please be more precise about what "up to the same permutation" means.
Thank you for your thoughtful feedback. We've addressed all your concerns below.
Q1: Lemma 2 Clarifications
- (i) Use or for Hadamard product?
- (ii) Clarify "same permutation"?
- (iii) (5) seems independent of —is this correct?
- (iv) What does "all hidden dimensions are active" mean?
- (v) Is it the exact or an upper bound on tunable parameters?
-
(i) Fixed.
-
(ii) "Up to the same permutation" means that the same permutation matrix is applied to all three parameters of the initial model . This yields a model that is functionally equivalent but with permuted hidden dimensions, giving Here, , , and are the parameters of before fine-tuning.
-
(iii) The quantity in (5) does, in fact, depend on the initial model through the constraint set . To prevent confusion, we rephrased Lemma 2 to highlight this constraint set.
-
(iv) It means that (for target model ), all hidden dimensions are non-zero.
-
(v) When the assumption (iv) holds, it refers to the exact minimum.
Preview Revised Lemma 2 with improved clarity (https://anonymous.4open.science/r/ce7/l2.png). Notations are from Sec. 5.1.
Q2: Theorem 1
- (i) Define and —are they from the target model?
- (ii) Does "selectively fine-tuning" imply a specific algorithm?
- (iii) Clarify "accurately represent."
- (i) and are indeed the number of layers and the hidden dimension of the target model We have updated statement for a better introduction of the target model and the corresponding notations.
- (ii) The statement focuses solely on model expressiveness, independent of training algorithms.
- (iii) It means "functionally equivalent," i.e., the fine-tuned model satisfies for all input sequences .
Preview Revised version of Theorem 1 with improved clarity (https://anonymous.4open.science/r/ce7/t1.png).
Q3: Lemma 1
- (i) Lemma 1 covers only expressiveness, not optimization—please comment.
- (ii) Lemma 1 suggests fine-tuning linear projections alone is sufficient; does this diminish your method’s significance?
- (i)
New Exp.Lemma 1 analyzes only theoretical expressiveness. Empirical results (https://anonymous.4open.science/r/ce7/f1.png) confirm fine-tuning alone matches tuning , , across three GLUE tasks. - (ii) Great point. To clarify, in Lemma 1, and are fixed. While 's expressiveness includes that of , , and , it does not cover the expressiveness of , which is essential for seq2seq operations. Thus, offers a portion of expressivity for SSM, yet it alone remains insufficient to attain optimal performance.
Preview Updated Lemma 1 to emphasize fixed (https://anonymous.4open.science/r/ce7/l1.png) and will further smooth the Sec. 4–5 transition.
Q4: Be clear whether the model is S4 or S4D.
We specifically use S4D (diagonal ) and will clarify this in Sec. 1, 3, and 5.
Q5: Fine-tuning strategies lack mathematical details.
Thanks. We'll revise Sec. 3 ("Preliminaries") to include mathematical details on fine-tuning methods.
Q6: What is the fine-tuning time compared to LoRA?
New Exp. Based on your feedback, we conducted two additional experiments: (i: https://anonymous.4open.science/r/ce7/f2.png) a performance comparison between SDT and LoRA under varying time budgets in a synthetic setting; and (ii: https://anonymous.4open.science/r/ce7/f3.png) a runtime analysis of SDT and LoRA for training on a single batch using a pretrained model, across different sequence lengths. We observe that our method is slightly more efficient than LoRA, particularly as the sequence length increases, because LoRA introduces additional matrix multiplications, while SDT does not.
Q7: How were , in Table 3/4 chosen, and why is it a fair comparison to LoRA?
New Exp We fix and sweep . To ensure a fair comparison with LoRA, we use similar parameter budgets and hyperparameter sets of the same size for both methods: 45 configurations per method (15 learning rates × 3 method-specific settings). For LoRA, we additionally compare the three chosen configurations against alternative settings with similar parameter budgets (https://anonymous.4open.science/r/ce7/f4.png), showing that our selected configurations perform reasonably well.
Final Note: In addition, we’ll summarize all notations in the appendix. Since you said raising your score would be “an easy reach” once issues were addressed, we hope our response meet that bar. We’d appreciate an updated score and are happy to clarify anything further.
This paper investigates the performance of popular parameter-efficient fine-tuning methods (PEFT) (e.g., LoRA and its variants) when applied to SSMs like Mamba and hybrid models such as Jamba. It finds that LoRA-based methods consistently outperform other PEFT approaches, especially when applied to linear projection matrices, but fail to improve performance significantly when directly applied to SSM modules. To address this, the authors propose Sparse Dimension Tuning (SDT), a specialized PEFT method designed explicitly for fine-tuning SSM modules by selectively training specific channel and state dimensions based on theoretical insights. Extensive experiments demonstrate that combining SDT for SSM modules with LoRA for linear projections achieves state-of-the-art performance across diverse tasks, including natural language understanding, generation, and computer vision benchmarks, confirming SDT's effectiveness and efficiency compared to existing methods.
给作者的问题
See above.
论据与证据
Yes. The claims are backed by extensive experiments and theoretical analysis.
方法与评估标准
The proposed methods and evaluation criteria make sense for the problem. The paper introduces SDT specifically to address the limitations of existing PEFT methods on SSM-based models, and it benchmarks these methods using a diverse set of tasks and datasets—including GLUE, DART, SAMSum, Spider, CIFAR-10, and CelebA—which are well-recognized for assessing performance in both language and vision applications.
理论论述
The proofs for the main theoretical claims, e.g., Lemma 1, Lemma 2, and Theorem 1 and their logical structures and derivations appear sound under the stated assumptions.
实验设计与分析
The authors benchmark PEFT methods on both synthetic datasets (using deep S4 models) and real-world tasks spanning GLUE, DART, SAMSum, Spider, CIFAR-10, and CelebA with SSM-based (Mamba) and hybrid (Jamba) architectures; while these experiments are well-controlled with careful hyperparameter tuning and fair parameter budget comparisons, a minor issue arise:
- The focus on SSM-based and hybrid models limits the generalizability of the findings, suggesting that broader architectural testing could further strengthen the evidence.
补充材料
I examined the extended related works on SSMs and PEFT, where Section C provided additional details on the experimental setup and extended benchmarking results, Section D offered in-depth proofs and theoretical analyses, and Section E presented expanded experimental evaluations on deep S4 models and additional results for Mamba and Jamba.
与现有文献的关系
The paper’s contributions are well-integrated with the broader scientific literature by extending established PEFT methods—such as prompt tuning, prefix-tuning, BitFit, and especially low-rank adaptations like LoRA—to the realm of SSMs, which have been previously developed for efficient long-sequence modeling.
遗漏的重要参考文献
No.
其他优缺点
Strengths:
- Creative integration of established PEFT techniques like LoRA with a novel Sparse Dimension Tuning (SDT) method tailored for SSM modules.
- Advances the state-of-the-art in fine-tuning SSM-based and hybrid models, which is valuable for efficient language modeling.
Weakness:
- Limited evaluation on a broader range of model architectures beyond SSM-based and hybrid models, which may affect the generalizability of the findings.
其他意见或建议
See above.
We thank the reviewer for the encouraging feedback, particularly for recognizing that (i) our method is creative and rational, (ii) our claims are supported by extensive experiments and theoretical analysis, (iii) our theoretical claims are sound, (iv) we advance SOTA results in fine-tuning SSM-based and hybrid models, and (v) our contributions are well-integrated with the broader scientific literature.
Q: Limited applicability beyond SSM-based and hybrid models.
Our proposed SDT algorithm is intentionally tailored for SSMs, with our theoretical analysis and experiments specifically focused on this model class.
Regarding your concern, while applying SDT to pure Transformers is beyond our current scope, we demonstrate its broader potential through experiments on Mamba-II, whose SSD module closely resembles attention [1]. When applied to SSD modules with LoRA on linear projection matrices, SDT consistently outperforms LoRA alone on Mamba-II across diverse tasks (Sec E.2).
Furthermore, we believe the theoretical insights and proof techniques from our analysis could prove valuable for developing PEFT methods for other architectures.
References:
[1] Dao, T. and Gu, A. Transformers are SSMs: Generalized models and efficient algorithms through structured state space duality. In International Conference on Machine Learning, 2024.
Final Note: Thank you again for your valuable comments. We are grateful that you appreciate our paper’s contributions in theory, methodology, and experimentation. If you have any remaining questions, please do not hesitate to let us know. Assuming our responses have addressed your concerns satisfactorily, we kindly ask you to consider raising your score and supporting our paper.
Thank you for the clarification. I will keep my current score.
We appreciate your engagement during the discussion, and thank you for supporting the acceptance of our paper.
This paper investigated the PEFT for SSM, like MAMBA.
给作者的问题
- The proposed PEFT is incremental, such idea is mainly from the PEFT for transformer-based LLM.
- The theoretical analysis is mainly for S4 or S6, how about S5 and Mamba2?
论据与证据
yes
方法与评估标准
yes
理论论述
yes
实验设计与分析
yes
补充材料
yes, the experimental parts
与现有文献的关系
it mainly focus on PEFT for SSM.
遗漏的重要参考文献
NO
其他优缺点
Strengths:
-
The paper is well written, and there are many PEFT studies for Transformer-based LLM, not many PEFT for SSM.
-
The paper contains many experiemntal result to support their claim in the supplementary materials.
Weaknesses:
- The proposed PEFT is incremental, such idea is mainly from the PEFT for transformer-based LLM.
- The theoretical analysis is mainly for S4 or S6, how about S5 and Mamba2?
其他意见或建议
No
We thank the reviewer for acknowledging that (i) our paper is well written, (ii) it tackles the gap where PEFT studies favor Transformer over SSM, and (iii) it contains many experimental results to support the claims.
Q1: The proposed PEFT is incremental, such idea is mainly from the PEFT for transformer-based LLM.
We respectfully disagree. SDT is not just an incremental tweak of PEFT for Transformer-based LLMs—it’s built on a fresh theoretical analysis tailored for SSMs’ unique parameter structure. Beyond that, our paper delivers substantial contributions that stand firm, even if the method is viewed as incremental by the reviewer:
- (i) (One of the First) A systematic benchmark of PEFT methods on SSM models,
- (ii) (First) A theoretical analysis of PEFT in the SSM setting, and
- (iii) (One of the First) A new PEFT method designed specifically for SSMs, supported by both theory and strong empirical results.
Together, these contributions offer new insights into PEFT beyond what has been explored in Transformer-based LLMs.
Q2: The theoretical analysis is mainly for S4 or S6, how about S5 and Mamba2?
-
New Theoretical ResultsS5: Although our original analysis primarily focused on S4 and S6, based on your comment, we have extended our theoretical results (Lemma 2 and Theorem 1) to include S5. Specifically, Lemma 2, which characterizes the minimal number of parameters required to update a frozen S4 model for functional equivalence to a target S4, has been extended to S5 with similar conclusions, accounting for multi-channel handling. Additionally, Theorem 1, regarding the expressive power of SDT-P with LoRA on simplified SSM-based models, also holds true for S5. We provide the detailed extension for S5 here: https://anonymous.4open.science/r/ce7/l9.png. -
Mamba-II: Although extending our theoretical analysis to Mamba-II is non-trivial, we successfully adapt our method to this architecture, as detailed in Sec. C.2 and E.2, and evaluate it on diverse tasks, including GLUE, DART, SAMSum, and Spider. Our method consistently outperforms LoRA alone on Mamba-II, highlighting its generalizability beyond the architectures explicitly covered in our theoretical analysis. We will acknowledge this theoretical limitation explicitly in the conclusion section, emphasizing that extending our analysis to Mamba-II is an important direction for future work.
Final Note: Thank you for sharing your concerns. While we understand your perspective, we kindly remind you of the substantial contributions our paper makes and the efforts we've undertaken to address your comments. We would greatly appreciate it if you could reconsider your evaluation score and support the acceptance of our paper.
This paper investigates how PEFT methods perform on State Space Models (SSMs) (e.g. the Mamba architecture) and identifies which model components are best to target. It provides a comprehensive benchmark of existing PEFT techniques on SSM-based language models and hybrid architectures with Jamba. A key finding is that LoRA and its variants consistently outperform other PEFT approaches on SSM models. However, LoRA’s benefit comes mainly from fine-tuning the linear projection matrices (e.g. input/output projections) – applying LoRA to the SSM-specific components does not yield additional gains. In fact, prompt-based tuning methods (like prefix or prompt tuning) are found largely ineffective on SSMs, essentially only adjusting the initial hidden state (a severely limited form of fine-tuning, as proven in the paper’s Proposition 1). These insights highlighted a gap: none of the standard PEFT methods are well-suited for the internal SSM parameters. To address this, the authors propose a new method called Sparse Dimension Tuning (SDT) – a PEFT strategy tailored for SSM layers. SDT works by selectively fine-tuning only a subset of the state-space channels and state dimensions, while freezing or pruning the rest. When combining SDT (for SSM layers) with LoRA (for the linear projection layers), the paper achieves state-of-the-art fine-tuning performance on multiple benchmarks. The approach matches or surpasses the best existing methods while training only a small fraction of model parameters, validating the effectiveness of the proposed methodology across extensive experiments.
给作者的问题
No.
论据与证据
Yes.
方法与评估标准
Yes.
理论论述
No.
实验设计与分析
Yes.
补充材料
Yes. Appendix E.
与现有文献的关系
It is related to Parameter Efficient Fine Tuning, State Space Models and Hybrid State Space Models. No previous works on systematic studies of PEFT on SSMs.
遗漏的重要参考文献
No.
其他优缺点
Stength:
- This is a well executed paper that systematically studied PEFT for the emerging neural architectures, including Mamba, Mamba 2 and its hybrid variant. The empirical results are strong and comprehensive.
其他意见或建议
No.
We are delighted that the reviewer likes the paper, recognizing it as (i) well-executed, (ii) a systematic study of PEFT for emerging neural architectures (Mamba, Mamba 2, and hybrid variants), and (iii) providing strong and comprehensive empirical results. Thank you for your encouragement!
This work performs a detailed study of parameter-efficient fine-tuning (PEFT) for the recently popular family of state space models (SSM). The contributions include a systematic empirical study of existing PEFT methods, and introducing a new method tailored for SSMs that applies to several variants including S4, Mamba, and more models that the authors included during the reviewer discussion. Reviewers agreed that the problem domain of PEFT for SSMs is important and relatively new, and that the paper's contributions are well-motivated and sound.