PaperHub
5.5
/10
Poster4 位审稿人
最低2最高4标准差0.7
2
3
3
4
ICML 2025

Logits are All We Need to Adapt Closed Models

OpenReviewPDF
提交: 2025-01-23更新: 2025-07-24
TL;DR

We propose a “Plugin” framework that uses token-logit reweighting at inference to adapt closed-source LLMs to new domains without retraining or accessing model weights.

摘要

关键词
Distribution ShiftBlack-box ModelReweighingDecodingLarge Language Models

评审与讨论

审稿意见
2

This paper studies the problem of adapting a black-box LLM to a downstream task, assuming access to the logits of output tokens. The authors propose a token-level probability reweighting algorithm that modifies token logits during inference. The core idea is to frame the adaptation problem as label noise correction in supervised classification and leverage an autoregressive probability reweighting model to estimate logits for the downstream task. Theoretical justifications are provided, and empirical studies on benchmark datasets show that the proposed approach outperforms several existing adaptation techniques.

给作者的问题

see weakness 3 above. Is there any comparison with other SOTA methods [1,2] for black-box LLM adaptation?

论据与证据

yes

方法与评估标准

yes

理论论述

yes

实验设计与分析

yes

补充材料

yes

与现有文献的关系

This paper consider the adaptation of black-box LLM to downstream task, which is quite broader in scientific literature

遗漏的重要参考文献

yes

其他优缺点

Framing the problem as transition matrix estimation in the label noise correction framework and handling the challenge of an extremely large label space using an autoregressive reweighting model is new and insightful to me. The theoretical analysis effectively establishes key properties of the proposed algorithm.

However, 1. the assumption of logit access significantly weakens the overall novelty and practical contribution of the work. The approach relies entirely on closed-source LLMs exposing logits, which is not currently supported by most commercial APIs, making the proposed learning setting somewhat artificial.

  1. The method requires training a separate reweighting model, introducing higher computational costs compared to simpler adaptation techniques like prompt tuning or in-context learning.

  2. The baseline comparisons (ICL-1, ICL-3) are relatively weak, as increasing the number of in-context demonstrations could likely improve their performance, making the reported advantage of the proposed method less conclusive. Meanwhile, other baseline adaptation method are not compared [1,2]

[1] Black-Box Tuning for Language-Model-as-a-Service. ICML 2022 [2] Black-box Prompt Learning for Pre-trained Language Models. TMLR

其他意见或建议

n/a

作者回复

We thank the reviewer for their feedback, which has helped us strengthen our paper.

Regarding Logit access assumption The central goal of this paper is to encourage closed-source LLM providers to offer logit-level access as a practical middle ground when releasing full model weights is not feasible due to IP or privacy concerns. Our theoretical and empirical results show that even with this limited access, effective domain and task adaptation is possible. Unlike prior work assuming either full white-box access or fully opaque models, we demonstrate that logit access enables fine-grained control without exposing proprietary internals. With this work, we aim to motivate commercial providers to adopt this feasible and impactful compromise, filling a key gap in current adaptation methods.

Computational Cost vs. Simpler Methods While prompt-based methods like prompt tuning or ICL may seem simpler, they often require extensive trial-and-error and show high variance, as reflected in our results. In contrast, our reweighting model offers consistent, theoretically grounded gains with lower overhead than exhaustive prompt engineering. It is also analogous to established methods like LoRA or Adapters in white-box settings, requiring a manageable training overhead. Ultimately, the stability, reliability, and performance improvements of our approach outweigh the modest compute required, especially when compared to the variability and manual tuning burden inherent in prompt-based techniques.

Baseline Comparisons: ICL Variants and API Access-based Methods Based on the reviewer’s suggestion, we extended ICL baselines to include 5, 8, and 10 examples, and implemented Diao et al. ([2], Black-box Prompt Learning, 2023) with the recommended 75 API calls as mentioned in the paper. We did not include Sun et al. ([1], Black-Box Tuning, 2022), as Diao et al. [2] already demonstrated superior performance.

Importantly, if logit access is available, Plugin can be layered on top of any prompt-based method using the best-found prompt. For instance, our Zeroshot prompt (noted in line 364, left column) is reused across all methods. We also apply Plugin on top of the best ICL variants (ICL-8/10) and Diao et al. [2]. The table below presents results across all datasets using GPT2-XL as the base model.

E2E NLGBLEURouge-1Rouge-2Rouge-LMETEORCIDErNIST
ICL-50.12260.43190.21940.30950.41720.31620.7281
ICL-80.15370.44320.24390.31800.42680.35590.8253
ICL-100.15820.44590.25020.32010.45280.41250.9015
Diao et al.0.22870.50240.28460.39220.46280.42160.8625
Plugin0.24700.55360.30840.42130.50570.54551.2736
ICL (best) + Plugin0.39410.67130.40270.53790.59230.61721.5472
Diao et al. + Plugin0.45270.71260.51260.60270.62140.70022.0817
WEB NLGBLEURouge-1Rouge-2Rouge-LMETEORCIDErNIST
ICL-50.08260.36250.17250.25170.32610.18260.2614
ICL-80.09430.38260.19260.28250.34250.20160.2611
ICL-100.08130.35280.17180.25420.33210.19060.2425
Diao et al.0.10240.40160.22430.30170.35270.43210.2631
Plugin0.16730.46160.25270.37570.38950.89870.2646
ICL (best) + Plugin0.19260.50260.27350.39270.38720.91230.4267
Diao et al. + Plugin0.21370.60260.30210.59280.57661.08260.6142
AdidasBLEURouge-1Rouge-2Rouge-LMETEORCIDErNIST
ICL-50.03450.26540.03930.16010.18630.03380.6856
ICL-80.04030.25270.04320.16280.18940.06150.6125
ICL-100.03820.25370.03250.15280.17250.04520.5926
Diao et al.0.04170.26150.06710.17100.18260.08610.6034
Plugin0.06000.27100.07220.17250.19950.11950.6375
ICL (best) + Plugin0.05910.27610.07540.17360.20470.12730.6415
Diao et al. + Plugin0.06230.27920.07730.17590.21480.13250.7024

We observed similar results on Common Gen and will include them in the second response due to space constraints in the rebuttal.

As shown, Plugin outperforms ICL even with 10 examples and surpasses Diao et al. (2023). While ICL’s performance plateaus with increasing examples—despite higher inference cost and variance—Plugin consistently offers greater accuracy and stability. Moreover, combining Plugin with the best ICL or Diao et al. setups yields further gains, highlighting the value of logit-level access in enhancing prompt-based methods.

审稿意见
3

The paper proposes logit reweighting to adapt closed-source LLMs for task-specific generation without accessing model weights. By learning an autoregressive transition matrix from task data, it adjusts token probabilities during inference to align outputs with target domains. Experiments show improved style/keyword compliance over zero-shot and instruction tuning, advocating logit accessibility as a practical adaptation pathway. The method bridges theoretical label shift correction with efficient LLM customization.

给作者的问题

No further questions.

论据与证据

The claims are partially supported: empirical results on style/keyword alignment validate performance gains, but theoretical guarantees (e.g., distribution alignment) rely on idealized assumptions (e.g., perfect transition matrix estimation) not fully verified in real-world noisy settings. Scalability claims for large vocabularies lack rigorous analysis of trade-offs with token pruning.

方法与评估标准

The methods (logit reweighting via transition matrices) align with the goal of adapting closed LLMs using limited access (logits only), and domain-specific metrics (style/keyword accuracy) suit tasks like product descriptions. However, human evaluation is notably absent for assessing output quality, and generalization tests are limited to narrow domains (e.g., one brand), leaving broader applicability under-explored.

理论论述

The theoretical claim of distribution alignment (via logit reweighting under ideal transition matrices) is logically consistent given the assumptions, but the proof sketch (as described in the summary) assumes noiseless task-specific data and perfect estimation of the transition matrix—conditions unlikely in practice. No convergence rates or sensitivity analysis for estimation errors are provided, and empirical results do not explicitly validate the theoretical bound (e.g., measuring distribution divergence post-adaptation).

实验设计与分析

The experimental design has validity in using domain-specific metrics (e.g., keyword accuracy for product descriptions), but human evaluation is missing, leaving output fluency/coherence unverified. Comparisons to baselines (zero-shot, instruction tuning) are reasonable, but scalability tests lack depth—token pruning’s impact on rare tokens is unstudied. Domain generalization is under-tested (e.g., single-brand data), raising concerns about broader applicability.

补充材料

Yes, I reviewed the algorithm descriptions in Appendix A, the assumptions in Appendix B, and the experimental details in Appendix C.

与现有文献的关系

The paper connects to label noise correction literature (e.g., noise transition matrices in supervised learning) by reframing LLM adaptation as correcting “noisy” general-purpose token distributions. It extends these ideas to autoregressive generation, differing from prior LLM adaptation (e.g., prompt tuning, soft prompts) by relying solely on logits, aligning with resource-efficient methods like light-weight finetuning but avoiding weight access. Theoretically, it bridges domain adaptation (e.g., label shift theory) with LLM customization, advancing closed-model adaptation paradigms.

遗漏的重要参考文献

The paper does not cite controlled text generation methods (e.g., FUDGE Yang & Klein, 2021), which also modify logits using auxiliary models for task alignment. Additionally, Qiu et al. also investigated how to dynamically adjust logits by learning temperature parameters to adapt to different tasks.

[1] Yang & Klein, FUDGE: Controlled Text Generation With Future Discriminators, In NAACL, 2021. [2] Qiu et al., To Cool or not to Cool? Temperature Network Meets Large Foundation Models via DRO. In ICML, 2024.

其他优缺点

Strengths: The paper’s originality lies in creatively bridging label noise correction with autoregressive LLM adaptation, a novel conceptual link that unlocks practical utility for closed models. Its emphasis on logit accessibility addresses a critical industry need (adapting proprietary models without weight access), offering significant real-world relevance. The method’s lightweight design (task-specific matrix learning) is a pragmatic strength.

Weaknesses: While innovative, the framing underplays overlaps with logit manipulation in controlled generation (e.g., FUDGE) and distillation. Broader claims about domain generalization lack empirical rigor (limited to narrow tasks/brands), and theoretical assumptions (perfect matrix estimation) are underexplored in practical noisy settings.

其他意见或建议

Suggestions:

  1. Human evaluation: Include user studies to validate output quality beyond automated metrics.
  2. Error analysis: Quantify how token pruning affects rare tokens or domain-specific terms.
  3. Comparison to logit-based methods: Explicitly contrast with FUDGE or distillation to clarify novelty.
作者回复

We appreciate the reviewer’s encouraging words and thoughtful feedback.

Regarding human evaluation and limited generalization tests We already conducted a human evaluation (line 366, details in Appendix C.7) where three evaluators compared Plugin and ICL-3 on 100 Adidas samples, with Plugin preferred in 81% of cases—directly supporting output quality.

While the Adidas dataset reflects a specialized domain shift, our study extends beyond a single brand. As shown in Section 7.1, we also evaluate on WEB NLG, E2E NLG, and CommonGen—each representing its own distribution shift relative to the black-box model’s pretraining data. In Section 7.3, we further explore adversarial shifts by testing Plugin on models with known biases (e.g., infrastructure in WEB NLG, male-related concepts in CommonGen), demonstrating Plugin’s broad applicability across diverse and challenging settings.

For most datasets, we follow standard practices from PEFT literature (Hu et al., 2021; 2023a) that rely on automated metrics comparing outputs to well-formed references to assess overall quality. This combination of human judgments (for Adidas) and standard metrics (for broader benchmarks and Adidas) provides both qualitative and quantitative evidence of the Plugin’s generalization capabilities.

Token Pruning We do not perform token pruning; instead, Plugin continuously reweights token probabilities at each decoding step without removing any tokens. This soft upweighing preserves vocabulary coverage and improves domain adaptation (see case study in line 408). For further clarification, we compare the total occurrences of the top-50 Adidas domain words in Plugin’s predictions versus the base model in the same case study. Across all samples, Plugin’s outputs include 25.6% occurrences of these words, while the baseline contains only 13.8%. We will add this in the paper.

Regarding FUDGE, Qiu et al. We do cite FUDGE in line 249 (right column). While FUDGE uses attribute-specific discriminators to control generation (e.g., formality), our method enables free-form domain adaptation via a single auxiliary model. We considered FUDGE as a baseline, but it requires one discriminator per predefined attribute—unsuitable for broad or evolving domain shifts, which are hard to define upfront.

TempNet in Qiu et al. learns a single temperature per input and uniformly scales logits during generation. In contrast, Plugin reweights logits at each timestep, enabling finer, context-sensitive adjustments. Additionally, Qiu et al.'s use of DRO involves an inner maximization loop, making it more computationally intensive than our efficient empirical risk minimization (ERM).

Nonetheless, we include a comparison on the E2E NLG dataset, adapting TempNet to use ERM (instead of DRO) with GPT2-XL. As it applies global scaling per prompt, it underperforms in tasks requiring localized adjustments. These results will be added to the final paper.

MethodBLEURouge-1Rouge-2Rouge-LMETEORCIDErNIST
TempNet0.13250.46420.25160.30210.41260.36270.8027
Plugin0.24700.55360.30840.42130.50570.54551.2736

Regarding Theoretical Claims and Convergence Rate Theorem 5.1 holds under mild, standard assumptions (5.1, 5.2, B.1) commonly used in convergence analyses (Frostig et al., 2015; Chaudhuri et al., 2015; Mukherjee et al., 2022). The noisy estimation of the transition matrix T_t(θ_;xi,xj,Ft1)T\_t({\theta}\_{\star};x_i,x_j,\mathcal{F}^{t-1}) can be understood in two ways: as direct estimation error in the matrix itself, or as error induced by the function f_It(θ_;xi,xj,Ft1)f\_{I_t}({\theta}\_{\star} ; x_i, x_j,\mathcal{F}^{t-1}) on which it depends (see Assumption 5.1).

We adopt the latter view, estimating fIt()f_{I_t}(\cdot) under a sequence of noisy autoregressive loss functions 1(θ),,t(θ):RVR\ell_1(\boldsymbol{\theta}), \ldots, \ell_t(\boldsymbol{\theta}) : \mathbb{R}^{|V|} \rightarrow \mathbb{R} (see Assumption 5.2). Under Assumption B.1—bounded gradients and Hessians of fItf_{I_t}—we show that the expected estimation error can be reduced and prove upper and lower bounds in terms of a problem-dependent quantity σt2\sigma_t^2, establishing a convergence rate of Ω(σt2/t)\Omega(\sigma_t^2 / t).

We hope this clarifies. Our novelty lies in combining techniques from Frostig et al. (2015), Chaudhuri et al. (2015), Mukherjee et al. (2022), and Patrini et al. (2017), and presenting the first finite-time convergence analysis for transition matrix estimation in this autoregressive noisy loss setting.

We acknowledge that real-world settings may add noise in mapping fItf_{I_t} to the transition matrix. While this could generalize our framework, incorporating it into our finite-time guarantee is left for future work. Regarding sensitivity analysis, our convergence rate depends on the variance-like term σt2\sigma_t^2; higher variance leads to slower convergence, naturally capturing sensitivity to estimation noise.

审稿意见
3

The key idea of the paper is to treat next-token prediction as a label noise correction problem, where discrepancies between the LLM’s broad training distribution and task-specific data are modeled through a transition matrix that reweights token probabilities during inference. The proposed Plugin model consists of a small reweighting network trained on limited task-specific data, which, when combined with the black-box LLM’s logits, effectively steers text generation towards the desired distribution. The authors provide theoretical guarantees showing that this probability reweighting approach converges to the target distribution with sufficient task data. Extensive experiments on multiple text generation benchmarks (E2E NLG, WebNLG, CommonGen, Adidas product descriptions) demonstrate that the Plugin model outperforms in-context learning and other adaptation methods, achieving better alignment with domain-specific content while requiring minimal computational resources compared to full fine-tuning. The results suggest that access to token logits could enable more powerful model customization, advocating for broader API-level exposure of logits in commercial LLMs.

给作者的问题

How many flops would the method save compared to vanilla LoRA?

论据与证据

The paper makes several claims regarding the effectiveness of the Plugin model for adapting black-box LLMs, and most of these claims are supported by theoretical justifications, empirical experiments, and comparative evaluations. The core claim—that token-level probability reweighting using logits alone is sufficient for effective task adaptation—is backed by a formal label noise correction framework, where the authors derive theoretical guarantees showing that the Plugin model can align token distributions with task-specific data under mild assumptions. Additionally, extensive experimental results on four datasets (E2E NLG, WebNLG, CommonGen, Adidas product descriptions) demonstrate that the Plugin model outperforms baselines such as zero-shot inference, in-context learning (ICL), and naive probability combination methods across multiple evaluation metrics (BLEU, ROUGE, METEOR, CIDEr, NIST). The ablation studies further support the claims by showing that model quality, reweighting complexity, and domain adaptation capabilities contribute to improved performance. However, some claims could benefit from stronger empirical validation. For instance, while the paper asserts that the Plugin model is effective under distribution shifts, the experiments focus on relatively constrained dataset modifications (e.g., filtering training data by entity types), and it remains unclear how well the approach generalizes to more extreme domain shifts or adversarial settings. Additionally, while the authors argue that their approach is computationally efficient, they do not provide a direct comparison of training or inference costs against alternative adaptation methods like parameter-efficient fine-tuning (LoRA, adapters), leaving room for further evidence on the trade-offs between performance gains and computational overhead. Nonetheless, the overall evidence presented is strong and well-aligned with the claims, making the Plugin model a compelling approach for logit-based adaptation of closed-source LLMs.

方法与评估标准

The proposed Plugin model and its evaluation criteria are well-aligned with the problem of adapting black-box LLMs without modifying model weights. The token-level probability reweighting framework is a reasonable method given the constraints of closed-source LLMs, and the formulation as a label noise correction problem provides a solid theoretical foundation. The authors evaluate the approach using four diverse text generation datasets—E2E NLG, WebNLG, CommonGen, and Adidas product descriptions—each representing different aspects of controlled text generation and domain adaptation.

理论论述

I didn't check the orrectness of any proofs for theoretical claims.

实验设计与分析

sound and well-structured, but there are a few areas where further validation or additional analysis could strengthen the claims. The authors conduct experiments across four datasets (E2E NLG, WebNLG, CommonGen, Adidas product descriptions) and compare their Plugin model to several strong baselines including zero-shot inference, in-context learning (ICL), and a weighted combination of model predictions. The use of seven standard NLG metrics (BLEU, ROUGE, METEOR, CIDEr, NIST) ensures a comprehensive evaluation of output quality. They also include a sound human evaluation. The Plugin model is benchmarked against ICL and naive probability reweighting, but not against LoRA, Adapters, or QLoRA, which might be feasible alternatives in cases where some access to model weights is possible.

补充材料

No, I didn't review the supplementary material.

与现有文献的关系

The paper situates itself within the broader literature on adapting large language models (LLMs) without access to model weights, drawing from areas such as prompt engineering, in-context learning (ICL), parameter-efficient fine-tuning (PEFT), label noise correction, and black-box model adaptation. The key contribution—reweighting token probabilities using logits as an alternative to full fine-tuning—builds upon prior work in label noise correction (Patrini et al., 2017) and autoregressive modeling, adapting these ideas to the language model decoding process. The formulation of next-token prediction as a noisy supervised classification problem is a novel connection that extends prior work on calibrating LLM outputs (Huang et al., 2024; Kapoor et al., 2024).

遗漏的重要参考文献

Important references are missing: On the Duality between Gradient Transformations and Adapters. Lucas Torroba-Hennigen, Hunter Lang, Han Guo, Yoon Kim.

Tuning Language Models by Proxy. Alisa Liu, Xiaochuang Han, Yizhong Wang, Yulia Tsvetkov, Yejin Choi, Noah A. Smith.

A Study on the Calibration of In-context Learning. Hanlin Zhang, Yi-Fan Zhang, Yaodong Yu, Dhruv Madeka, Dean Foster, Eric Xing, Hima Lakkaraju, Sham Kakade

其他优缺点

The paper title is a bit bold and the methodology presented in the paper is conceptually interesting but complex, as it involves multiple layers of statistical modeling and adaptation that may not be immediately intuitive. The core idea—reweighting token probabilities using a transition model learned from task-specific data—relies on a label noise correction framework, which is commonly used in supervised classification problems but less so in language model decoding.

其他意见或建议

The Plugin model is positioned as an alternative to fine-tuning, but it is not directly compared to parameter-efficient fine-tuning methods such as LoRA, Adapters, or QLoRA. Including a discussion on when logit reweighting is preferable to PEFT would improve clarity.

作者回复

We thank the reviewer for the positive and insightful review.

Regarding Comparison of Plugin with Parameter Efficient Fine-tuning (PEFT) methods like LoRA, Adapters:

"... The Plugin model is positioned as an alternative to fine-tuning, but it is not directly compared to parameter-efficient fine-tuning methods such as LoRA, Adapters, or QLoRA..." "... How many flops would the method save compared to vanilla LoRA?"

We rely on only output logits of the black-box model and do not have access to its internal weights or architecture. Consequently, any form of fine-tuning—including parameter-efficient approaches like LoRA or Adapters—cannot be applied, which we clarify in the Introduction (lines 21–33, right column) and the Related Work (lines 261–273, left column). If we did have access to the model weights, parameter-efficient fine-tuning methods would indeed be the natural choice, as they use more information than just logits and are expected to yield better performance. Thus, we emphasize that the Plugin model is not an alternative to fine-tuning, but rather an approach that uniquely stands for adapting black-box LLMs which only provide logit access.

Nevertheless, to address the reviewer’s point, we conducted a comparison on the E2E NLG dataset by adding rank-𝑟=8𝑟=8 LoRA matrices in the QQ and VV attention layers of GPT2-XL. The results, which will be included in our final version, show that LoRA only slightly outperforms our Plugin in terms of task metrics. LoRA adds 2.46M parameters (rank-8, Q/V only), while Plugin adds 30.72M (one full layer). During inference (up to 64 tokens), LoRA requires 188.8B FLOPs while Plugin needs 196.2B FLOPs - a negligible difference in computational cost. Notably, the parameter and efficiency gap between LoRA and Plugin narrows when increasing LoRA's rank (r) while reducing Plugin's hidden dimensions - demonstrating how both approaches can be adaptively tuned to meet specific resource constraints while maintaining competitive performance. We would like to highlight that the fundamental distinction remains - LoRA requires full model access to modify internal layers, while Plugin enables post-hoc deployment without retraining.

E2E NLGBLEURouge-1Rouge-2Rouge-LMETEORCIDErNIST
Zeroshot0.05620.40130.16360.28620.36970.01870.5338
Plugin0.24700.55360.30840.42130.50570.54551.2736
PEFT (LoRA r=8)0.25170.57120.30790.43170.51620.52251.2172

Regarding Missing References

On the Duality between Gradient Transformations and Adapters. Lucas Torroba-Hennigen et al. 2025 Tuning Language Models by Proxy. Liu et al. 2024 A Study on the Calibration of In-context Learning. Zhang et al. 2023

Thank you for pointing out these relevant works. We will add them to our final version. In particular, we note that Liu et al. (2024) describes the same method introduced in Liu et al. (2021), which we already cite in line 252. Below, we clarify how our approach differs from each reference:

Torroba-Hennigen et al. (2025): They examine the equivalence between gradient transformations and adapters for efficient model adaptation, relying on full access to model weights and gradients. In contrast, our method requires no access to model internals; we adapt black-box LLMs solely by reweighting token-level logits.

Liu et al. (2024): As noted, we already cite their earlier work (Liu et al. 2021), which presents the same core idea of combining logits. Our WeightedComb baseline (line 252, right column) is directly inspired by their approach.

Zhang et al. (2023): We will add this reference to our discussion on calibration. Their study focuses on aligning model confidence with predictive accuracy by adjusting confidence scores as one increases shots in the few-shot learning setting. Unlike them, we explicitly modify the token predictions themselves, rather than just calibrating confidence.

Regarding Domain Shifts, Extreme Domain Shifts, and Adversarial Domain Shifts

We assume our black-box LLM already encodes extensive world knowledge. In this sense, any adaptation to a domain-specific dataset amounts to handling a distribution shift, as demonstrated in Section 7.1 with WEB NLG, E2E NLG, and CommonGen. Among these, the Adidas dataset represents a more extreme domain shift, given its specialized style of product descriptions. Additionally, the experiments in Section 7.3 can be seen as adversarial, since the Plugin is applied atop a model with known biases—for instance, the tendency to focus on infrastructure-related concepts in WEB NLG and male-related concepts in CommonGen. Although we did not explicitly label these settings “extreme” or “adversarial,” they do indeed meet those criteria to some extent, and we will clarify that in the final version.

We would also welcome any further examples or clarifications on what the reviewer would consider to be more extreme or adversarial distribution shifts in this context.

审稿意见
4

The paper tackles the issue of having to rely on prompt engineering while adapting closed-source LLMs. The proposed work formulates the problem as a supervised learning, where few task specific dataset is used to train the model. The closed source model is assumed to learn noisy labels of specific application, and by accessing logits of the tokens, the proposed loss function corrects the noisy label to adapt to a task. Extensive experiments are conducted on models, datasets, and with multiple evaluation criteria.

给作者的问题

What are the implications of making the transition matrix diagonal?

论据与证据

No issues found

方法与评估标准

No issues found

理论论述

No issues found

实验设计与分析

No issues found

补充材料

Yes. Experimental details.

与现有文献的关系

It is related to the adaptation of black-box LLM.

遗漏的重要参考文献

No issues found

其他优缺点

Strengths

  • The problem is well-motivated.

  • The approach to solving the adaption problem is novel. With this method, it can be used as a plugin to adapt a closed-source model.

  • Experiments are extensive to confirm the verification of the proposed method.

  • Theoretical analysis of the proposed method makes it guaranteed.

Weakness

Inclusion of some other closed-source models would make the experiment more comprehensive.

其他意见或建议

Please refer to previous sections.

作者回复

We are grateful for the reviewer’s positive remarks and valuable insights.

Regarding Inclusion of Closed Source Models We acknowledge that including more closed-source models could further strengthen the generality of our findings. However, most proprietary models currently do not expose their output logits. Hence, we could not experiment with them. Our approach specifically hinges on logit-level access—without revealing internal weights or architecture—which we believe is a comparatively straightforward adjustment for closed-source providers to implement, especially when compared to the complexities of releasing the full model.

By highlighting this limitation, we aim to encourage closed-source developers to consider offering logit-level access in the future, enabling more flexible and efficient adaptation methods for end-users.

Implications of making the transition matrix diagonal

Benefits

  1. Reduced Complexity: Learning a diagonal matrix involves only V|V| parameters, compared to V×V|V|\times |V| parameters in a full matrix. This makes training computationally more tractable.

  2. Straightforward Integration: Because the transition matrix is treated like a single vector, standard autoregressive models (e.g., GPT-2, LLaMA) can be used directly, scaling easily with the dataset size.

  3. Symmetric (Class-Independent) Noise: A diagonal assumption naturally corresponds to label flips occurring with equal probability among all incorrect classes. This aligns neatly with widely studied symmetrical label-noise scenarios.

  4. Stability in Estimation: Each diagonal entry is an autoregressive parameter independently learnt without depending on other classes (vocabulary), reducing the risk of overfitting and simplifying parameter estimation compared to a fully dense matrix.

Limitation

Although adopting a diagonal transition matrix reduces complexity and simplifies parameter estimation, it also prevents the model from capturing any off-diagonal confusions—that is, instances where certain tokens are more likely to be misclassified as particular other tokens. Real-world data may involve non-symmetric noise patterns, domain-specific synonyms, and context-dependent misclassifications. A purely diagonal matrix may therefore oversimplify these nuances, leading to diminished expressive power and potentially missing important relationships within the noise structure.

Overall, the diagonal constraint provides a practical, stable, and computationally efficient solution for autoregressive reweighing learning—one that is well-aligned with symmetrical label-noise setups—while sacrificing some flexibility in capturing sophisticated non-symmetric noise or domain-specific synonym patterns.

We plan to investigate non-diagonal structures to fully model these nuanced noise patterns and further improve adaptation in the future work.

审稿人评论

Thank you for your comprehensive reply to my questions.

I maintain my rating as accept.

最终决定

This paper presents a timely and well-motivated contribution to the adaptation of closed-source large language models (LLMs). The key insight is that access to token-level logits, if made available, would enable powerful post-hoc alignment mechanisms beyond prompt tuning. Framing next-token prediction as a supervised classification problem under label noise, the authors introduce a theoretically grounded reweighting framework, Plugin, which can steer black-box LLMs toward task-specific behavior using only logits and a small amount of labeled data.

The reviewers appreciated the novelty and practicality of the proposal, particularly its clean theoretical framing, empirical rigor, and relevance in the context of real-world LLM deployment constraints. Given its conceptual clarity, soundness, and strong empirical results, this paper makes a valuable contribution to the growing literature on black-box model alignment and LLM accessibility.