Understanding Multimodal LLMs Under Distribution Shifts: An Information-Theoretic Approach
We propose the first formal framework for characterizing the performance gap of multimodal LLMs under distribution shifts using information-theoretic metrics.
摘要
评审与讨论
This paper introduces an information-theoretic framework to analyze and understand the performance of Multimodal Large Language Models (MLLMs) under distribution shifts, which occur when the evaluation data differs from the instruction tuning distribution. The authors propose the concept of Effective Mutual Information (EMI), a metric that quantifies the relevance between input queries and model responses. EMI provides a more robust and theoretically grounded alternative to existing empirical metrics like the win rate. The paper derives upper bounds for the EMI difference between in-distribution (ID) and out-of-distribution (OOD) data, linking it to visual and textual distributional discrepancies. Through extensive experiments across 61 real-world distribution shift scenarios, the authors validate their framework, demonstrating strong correlations between EMI and model performance. The results suggest that EMI can be used for more reliable and cost-effective evaluation of MLLMs, particularly in situations where traditional evaluation methods may be computationally expensive or less transparent.
给作者的问题
Please refer to the weakness part.
论据与证据
Based on the results in Figure 1, the authors claim that as the severity of the shift increases, the performance degradation becomes more significant. First, part of the result is covered by the legend in text shift LlaVA v1.5, which requires some alteration. Second, the result in text shift does not clearly verify the statement. According to the context, the x-axis is sorted by severity of shifts, but the win rate in text shift isn’t decreasing monotonously. Also, LlaVA NeXT 7B performs better in Korean than in Arbaic, yet LlaVA NeXT 13B shows opposite performance respectively.
方法与评估标准
Yes.
理论论述
Yes.
实验设计与分析
The results in terms of Spearman correlation analysis and Kendall’s tau analysis needs further explanation, it is beneficial to specify the meaning of the value for better understanding. In Figure 3, all four model results are shown in one graph, making it difficult to understand the authors’ implications. Also, from the figure, it seems that EMID is a lot smaller than EMID upper bound, for example, in synthetic shift graph, EMID ranges from -0.02 to 0.10, but the upper bound ranges from 1.1 to 1.8. If not misunderstood, this shows that the upper bound is very loose and has little constraint on EMID.
补充材料
The appendix includes further description about EMI, implementation detail, additional experiments and proofs of the theory.
与现有文献的关系
This paper introduces Effective Mutual Information (EMI) to evaluate MLLMs under distribution shifts, extending information-theoretic frameworks to multimodal models. It provides a formal method to quantify performance degradation, addressing gaps in out-of-distribution generalization and model robustness in real-world applications.
遗漏的重要参考文献
No.
其他优缺点
Strengths
- Understanding MLLMs under distribution shifts seems a critical research problem.
- The paper is well constructed and easy to read.
Weaknesses
- In Figure 1, the authors claim that as shift severity increases, performance degradation worsens. However, part of the result in the text shift is obscured by the legend for LLaVA v1.5, requiring adjustment. Additionally, the text shift results do not fully support this claim, as the win rate does not decrease monotonically with shift severity. For instance, LLaVA NeXT 7B performs better in Korean than Arabic, while LLaVA NeXT 13B shows the opposite trend.
- The results from the Spearman correlation analysis and Kendall’s tau analysis require further clarification. It would be helpful to specify the meaning of the correlation values to aid in better understanding their implications.
- In Figure 3, combining all four models in one graph makes it hard to grasp the authors’ implications. Additionally, the EMID appears much smaller than its upper bound—e.g., in the synthetic shift graph, EMID ranges from -0.02 to 0.10, while the upper bound spans from 1.1 to 1.8. If not misunderstood, this suggests the upper bound is very loose and offers little constraint on EMID.
其他意见或建议
No.
A1. Legend issue in Fig. 1. and non-monotonic performance trend in text shifts.
Thank you for pointing out the visualization issue—we will revise the legend in Fig. 1 to improve clarity and avoid confusion!
- Regarding the non-monotonic trend in win rate under text shifts, this behavior arises in part from the inherent stochasticity of win rate computation based on GPT-4 API evaluations, making it fundamentally difficult to observe strictly monotonic trends in practice [1]. Additionally, the x-axis in Fig. 1 is sorted by embedding space distances—computed using CLIP ViT for visual shifts and XLM-RoBERTa for text shifts—which may not always reflect the true degree of distributional shift.
- That said, we still observe a meaningful overall relationship between embedding distance and performance degradation, both in the 27 natural shifts presented in Fig. 1 and in the 34 synthetic shifts shown in Fig. 6 of the Appendix. By taking inspiration from this empirical analysis, we derive a much more rigorous framework, i.e., EMID upper bound (Theorem 4.5), to quantify the performance gap that consistently shows statistical significance across diverse settings.
A2. Clarification for the meaning of Spearman correlation and Kendall’s tau.
- Spearman's and Kendall’s are both representative measures for monotonic relationships between two variables, where the former is preferred for detecting weak correlation and the latter is preferred to capture strong correlation in small sample sizes and is more robust to outliers with large sample sizes. Both of them are standard approaches to measure the correlation between LLM Judge score and other metrics [2], and are good for measuring the relation of two variables, even if they have different data types, e.g., discrete (win rate) versus continuous (EMI).
- Both correlation coefficients range from -1.0 (negative correlation) to 1.0 (positive correlation), where 0.0 indicates there is no monotonic relationship between two variables. For Spearman's , a 0.2-0.4 range of values denotes weak correlation, 0.4-0.6 denotes moderate correlation, and 0.6-0.8 and 0.8-1.0 denote strong and very strong correlations, respectively, and Kendall's can be similarly interpreted by multiplying 1.5, i.e., , to compensate its relatively smaller scale in practice.
- Our analysis in the paper (Table 2) indicates that the EMI consistently shows moderate or strong correlation with the LLM-judge evaluation metric, win rate, across different types of shifts and model architectures.
We will add this description in our future version of the manuscript. Thank you for the suggestion.
A3. Intricate visualization of Figure 3 and the tightness of EMID upper bound.
- On the left two panels in Figure 3, we presented the overall relationship between EMID and its upper bound, whereas we distinguished four different models on the right two panels to show the model-dependent difference over the relationship. We did so to show that Theorem 4.5 actually differentiates the model itself through the output entropy of each model , i.e., LLaVA NeXT shows higher sensitivity to the shifts compared with LLaVA v1.5 implied by the larger slope.
- In this work, we do not claim the tightness of the derived upper bound. Moreover, verifying the tightness of the proposed bound can be affected by the choice of estimators for the MI and JSD terms during empirical validation. However, we would like to emphasize that the bound shows consistent correlation with statistical significance across 61 cases of distribution shift over four different models, which means that our analytic bound of EMID effectively stands for performance degradation of MLLM. We appreciate your valuable concern, and we will explore devising a much tighter bound in future work.
Reference
- Judging LLM-as-a-Judge with MT-Bench and Chatbot Arena, Zheng et al. 2023
- PROMETHEUS: INDUCING FINE-GRAINED EVALUATION CAPABILITY IN LANGUAGE MODELS, Kim et al. 2024
Thank you for your detailed response. I have also read the comments from other reviewers. Most of my concerns have been adequately addressed. As a result, I would like to keep my score as '3', leading to acceptance.
Dear Reviewer vm1D,
The authors would like to express their sincere gratitude to Reviewer vm1D for taking valuable time to read and provide helpful feedback that contributes to improving the quality of this work! We are happy to hear that most of the concerns were adequately addressed, and thanks for acknowledging our contribution.
Best regards, The authors
This paper introduces Effective Mutual Information (EMI), a novel metric for quantifying the relevance between input queries and model responses in MLLMs. Unlike standard Mutual Information, EMI removes domain-dependent components, making it a more generalizable measure especially in OOD settings. The authors establish theoretical connections between EMI and win rate, providing an intuitive explanation of EMI’s effectiveness. Additionally, they derive an upper bound for the performance gap in OOD scenarios using EMID. Extensive experiments validate EMI’s utility, confirming its theoretical predictions and demonstrating its effectiveness across diverse distribution shift scenarios.
给作者的问题
-
The paper states, “In Eq. (4), we show that the autoregressive objective for instruction tuning (Eq. (1)) effectively estimates the lower bound of MI when the model’s representation capacity is sufficiently high.” However, while the term 𝛿 can be omitted under this assumption, I suppose that 𝐻(𝑃𝑌) could be significantly large, making Eq. (4) an inaccurate estimate of the MI lower bound.
-
The computation of EMI relies on pre-trained encoders (e.g., CLIP and XLM-R) for feature extraction, but the paper does not discuss the sensitivity of these encoders to domain shifts. For instance, CLIP may underperform on medical images, leading to distorted EMI estimates.
论据与证据
yes
方法与评估标准
yes
理论论述
yes
实验设计与分析
yes
补充材料
yes
与现有文献的关系
above average
遗漏的重要参考文献
n/a
其他优缺点
Strengths:
- This paper provides a theoretical framework to analyze and understand the impact of distribution shifts on MLLM performance via EMI.
- The intuition of this new metric EMI is explained via analogy with excess risk, effective robustness and win rate.
- The empirical findings strongly support most of the theoretical conclusions.
Weaknesses:
- I appreciate the theoretical part. However, the current theories do not address how modality fusion and modality interaction affect generalization, despite these techniques being commonly employed in (MLLMs).
- In Theorem 4.4, the paper assume that the -representation capacity holds, where itself requires to be quantified, e.g. it it closed related to the model size, the number of training samples and the input dimension. Given that the assumption fundementally underpins the theoretical contributions presented, I suggest expanding the discourse to elucidate the quantitative interdependencies between -representation capacity and these factors.
- The paper does not demonstrate how to leverage EMI or the EMID upper bound to guide model optimization (e.g., designing robust training objectives or adaptation strategies). It only uses EMI as a post-hoc evaluation tool.
其他意见或建议
n/a
A1. Effect of modality fusion/interaction on generalization.
- MLLMs commonly undergo a modality alignment phase during training, which may affect generalization, and it is known that modality fusion can reduce the sample complexity to improve generalization [1]!
- As noted in
line 250-254in our paper, can be factorized into where the conditional MI terms and encapsulate modality interaction. - Based on this factorization, we can define per-modality EMI based on and , and then derive a new upper bound that is constructed with and terms to capture the effect of modality interaction. We leave the explicit derivation for future work.
A2. Interdependencies between -representation capacity and the model size, number of training samples, and input dimension.
The -representation capacity assumption captures the minimum achievable discrepancy between truth distribution and the model's distribution . Due to the expectation and operator, it does not depend on the training sample size but is mainly influenced by model capability.
Specifically, as the models become more expressive—e.g., by increasing model size [5] and leveraging advanced positional encoding [6]—MLLM approaches the universal approximator of sequence-to-sequence mapping [6,7], as a result, the minimum expected discrepancy tends to decrease, leading to a smaller .
We will elucidate this in the next version, thanks!
A3. How to leverage EMID upper bound to guide model optimization?
While our primary focus is on presenting the theoretical framework to quantify the performance gap of MLLMs, we also showcase an application of the EMID upper bound in this rebuttal, a regularization term for visual instruction tuning.
Due to the space limit, we have included the setups and results in the rebuttal to reviewer o1Fb, response A3. Please refer to that thread! As shown in the tables, our instantiation of Theorem 4.5 can indeed be used to optimize model to improve robustness under shifts.
A4. Eq. (4) can be an inaccurate estimate of MI lower bound due to potentially large .
- In the Eq. (4): , maximizing through instruction tuning can be interpreted to learn a parameter that maximizes rather than solely .
- We do not claim that the log-likelihood term is a tight lower bound of MI but rather suggest that maximizing can implicitly maximize the MI between input and model's response. We will revise the paper to make this clear.
- To validate this, we reproduce the visual instruction tuning of LLaVA-v1.5-7B on a 10% subset of data, and show how the empirical estimate [2] of MI evolves during training.
Step 1 0.166 5 0.172 20 0.182 100 0.187 200 0.194 500 0.197 - As shown above, visual instruction tuning can effectively maximize MI between input and model response.
A5. Encoder sensitivity analysis to domain shifts for EMI estimation.
- In Table 4 and 5 of Appendix, we already discussed two alternative choices of the encoders, e.g., [3], and showed that our theoretical claims hold in practice consistently across varying encoders with statistical significance.
- We further conduct encoder sensitivity analysis under domain shifts by replicating experiments on medical domains with a CLIP-ViT-B32 and XLM-RoBERTa encoders.
- Specifically, we use 200 samples of LLaVA-Med [4], get three splits of them based on embedding distance with COCO images, and translate English queries into six different languages used in the paper by using GPT-4o to induce 28 subsets of shifts to conduct correlation analysis for EMID and its upper bound.
Model Pearson -val LLaVA-v1.5.7B 0.93 0.00 - We see the correlation between EMID and its upper bound estimates is very strong, even though the medical image and text are relatively minority instances compared with general object and text, implying that our theorem robustly holds even on the special domains that encoders may not excel at.
Reference
- A Theory of Multimodal Learning, Lu 2023
- A Contrastive Log-ratio Upper Bound of Mutual Information, Cheng et al. 2020
- Universal Embeddings with Multimodal Large Language Models, Jiang et al. 2024
- LLaVA-Med: Training a Large Language-and-Vision Assistant for Biomedicine in One Day, Li et al. 2023
- Scaling Laws for Neural Language Models, Kaplan et al. 2020
- Your Transformer May Not be as Powerful as You Expect, Luo et al. 2022
- Transformers are Universal In-context Learners, Furuya et al. 2024
I struggled to understand the paper so take this with a grain of salt:
The authors suggest there is a risk involved in using multimodal models (language models conditioned on visual input) and suggest a way to measure this is using a difference of mutual information between a model and the distribution of the training data (is this right?) – though it remains unclear to me how that distribution is gotten.
给作者的问题
I think the biggest problem with the paper is the writing and underspecified mathematics. The text is unclear throughout the abstract, introduction, and the motivation of what is being done. I don't find it clear what problem is being solved, nor what the exact configuration in which your method applies is.
More concretely, I don’t find the distributions and the distribution shifts well-defined making it very hard to follow any derivations or reasoning. Even if these (in your mind) are somewhat standard conditional factorizations, I want to see it spelled out, for every such , otherwise it’s very hard to be sure that what is written is correct. It was also unclear what is trained and what is not trained.
The authors define mutual information as a function of a single distribution with an implicit factorization instead of over two random variables. This did not make it easier to follow the writing. For instance, in equation 7 of the EMI (a main contribution), what is the definition of , the definition of is hardcoded for a distribution based on the factorization given. If there is some marginalization used to define it this needs to be defined, or maybe stick to the mutual information between two random variables.
论据与证据
The abstract talks about the risk of MLLMs and safe and reliable applications, and the proposed framework about quantifying risk under distribution shifts. It’s not clear what this risk is. Would it make sense to just call it the performance?
You talk about multimodal language models in general in the introduction, but it seems this is specifically about (bi-modal) vision and language models with vision only in the input. It could also be made clearer that the paper is only about textual output.
I also don’t see the reason for focusing on vision in the input and instruction tuning. It seems to me the part of the ideas I followed apply to sequence models in general. As such the question is to what degree a conditional distribution (given some prefix) differs from the non-conditional distribution.
方法与评估标准
I don’t understand how the various probability distributions are defined. I don't understand the definitions of the visual and textual shifts, e.g. the relation between and . How are these defined exactly in terms of next token probabilities? What is and ? You only define KL with some arbitrary distribution P.
How are and defined? Are these trained models? Are all the same trained model differing only in prefixes? E.g. as used in equation 3. For all of these, I would expect to see exact definitions of the distributions in terms of the exact model configuration you say your method works for.
理论论述
I struggled to follow the definitions of the authors and so did not go through the proofs.
实验设计与分析
No, since I struggled to understand what is being done
补充材料
No
与现有文献的关系
Unclear.
遗漏的重要参考文献
Unclear
其他优缺点
See other answers
其他意见或建议
Some notes:
In the motivation with visual shifts, I would put the definitions of the variables before the shift examples, it was confusing.
Random variables: Could you define the domains of the random variables in line 93? How is (X_v, X_t) combined into a single sequence?
Line 101: instruction tuning has not been introduced. Line 99: joint population → joint probability? Equation 1, should this not be argmin? Equation 2. Should this be P_{\theta} instead of \theta
We appreciate sTSP's effort to read our paper and provide comments. Here is a notation table and our responses.
| Var. | Def. |
|---|---|
| where | a random variable (r.v.) of a text input sequence with length of tokens in vocabulary |
| where | a r.v. of a -dimensional image embedding sequence with length of tokens produced by a frozen vision encoder |
| a joint r.v. of an input query constructed with a tuple of and | |
| where | a r.v. of a text response with length of tokens |
| a probability distribution (p.d.) of an input | |
| a p.d. of a visual input | |
| a p.d. of a text input | |
| a p.d. of a text response | |
| a conditional p.d. of a text response given input | |
| a joint p.d. of input and response | |
| model's prediction p.d. for a response given input |
What problem is being solved? What is the risk?
- As we noted in the introduction, MLLMs suffer from performance degradation when they encounter distribution shifts. We denote risk as this performance degradation.
Problem focus: As noted in the abstract, introduction, and motivation sections, our goal is to quantify the performance degradation of MLLMs under distribution shifts by presenting an information-theoretic framework.
It seems this is about (bi-modal) vision and language models with vision only in the input.
- As we noted in
L107-109, MLLMs take multimodal input (both visual and text) to produce text output, not "vision only in the input". - The term MLLM is commonly used in the literature to denote LLMs that receive a visual input as well as text [1,2], so we took this term by following the convention.
Definition of .
- In
L92-103, we put the definition of random variables and distributions. denote the probability distributions of the input , target response , and their joint , respectively. - The is a model being trained.
Definitions of and . How are these defined in terms of next token probabilities? What is and ?
As noted in L96-97, and are the sequences of visual and text input tokens, so the and are the corresponding distributions. and are defined at the input, which should not be confused with the next token probabilities (at the output level).
Put the definitions of variables before shift examples. Domains of the variables in L93? How is (X_v,X_t) combined into a sequence?
- We put the definition of all variables in
L92-103before the motivation section. - We concatenate the visual input tokens and textual input tokens into a single sequence, where the visual tokens are obtained by encoding an image using a vision encoder (CLIP-ViT), and then projecting them into the language embedding space. This follows the standard practice in MLLMs, where visual tokens are prepended to the text tokens to form a unified input sequence.
Definition of MI and EMI
- Our main interest is to express the model performance gap across different distributions. Thus, instead of defining MI with individual variables, we define MI as their joint distribution, which is mathematically equivalent to random variable-based MI, as can be seen in Eq 3.
- The definition of EMI was explicitly introduced in Eq 6. Please refer to
L168-169.
L101: instruction tuning has not been introduced. L99: joint probability? Eq 1, should this not be argmin? Eq 2. Should this be P_{\theta} instead of \theta
- Instruction tuning was introduced from
L104. - In statistics, the population (distribution) [3] is used to denote a distribution of the entire collection of objects in contrast to a sampled distribution.
- In Eq 1, both min and argmin can be valid, where the former aims at the objective-centric whereas the latter stands for the parameter-centric perspective.
- In Eq 2, we use the first argument to denote the data distribution that the metric is computed on, and use the second argument to denote the model parameter to be evaluated. We will use in the revised version.
Reference
- A Fully Open, Vision-Centric Exploration of Multimodal LLMs, Tong et al. 2024
- Exploring The Design Space for Multimodal LLMs with Mixture of Encoders, Shi et al. 2025
- Sampling of Populations: Methods and Applications, Levy and Lemeshow 2013
The paper proposes an information-theoretic framework to analyze the performance of multimodal large language models (MLLMs) under distribution shifts. It introduces Effective Mutual Information (EMI) to quantify the relevance between input queries and model responses. The authors also derive an upper bound for the EMI difference between in-distribution (ID) and out-of-distribution (OOD) data, connecting it to visual and textual distributional discrepancies. Extensive experiments on real benchmark datasets validate the theoretical insights.
给作者的问题
- How do the assumptions made in the theoretical framework impact the applicability of the results to more complex real-world scenarios? Can these assumptions be relaxed in future work?
- The paper mentions the potential use of the upper bound of EMID as a regularizer during post-training or test-time adaptation. Can you provide more details on how this could be implemented and its potential impact on model robustness?
论据与证据
The claims made in the paper are supported by clear and convincing evidence.
方法与评估标准
Yes
理论论述
The correctness of the theoretical claims, including Lemma 4.3, Theorem 4.4, Theorem 4.5, and Theorem 4.6, is supported by detailed proofs in the supplementary material. The derivations appear to be sound.
实验设计与分析
All checked
补充材料
Implementation Details and D.1. Proof for the relationship between EMI and preference model are reviewed.
与现有文献的关系
The proposed information-theoretic framework for analyzing MLLMs under distribution shifts provided insights for better understanding and alleviating that issue.
遗漏的重要参考文献
N/A
其他优缺点
- Strengths: The connection between EMI and win rate provides a practical and efficient alternative for model evaluation.
- Weaknesses: Evaluations are only made on LLaVA v1.5 and LLaVA NeXT. It could benefit from involving the SOTA and representative MLLMs like Qwen2.5-VL and InternVL2.5.
其他意见或建议
N/A
A1. Applicability on SOTA MLLMs.
Thanks for the suggestion! Following your comment, we additionally conduct the full evaluation withQwen2.5-VL-7B-Instruct and InternVL2.5-7B.
Specifically, we first evaluate the official release of Qwen2.5-VL-7B-Instruct and InternVL2.5-7B models on 35 synthetic shifts, then compute EMI estimates over the model responses. We perform a correlation analysis between estimates of EMI difference (EMID) and its upper bound. Consistent with our existing finding, we observe a strong correlation between EMID and its theoretical upper bound in both models.
| Model | Pearson | -val |
|---|---|---|
| InternVL2.5-7B | 0.67 | 0.00 |
| Qwen2.5-VL-7B | 0.81 | 0.00 |
A2. How do assumptions impact the applicability of proposals to more complex cases? This is an insightful question. First, to claim the closeness between EMI and win rate (Thm 4.4), we assumed -representation capacity of the MLLM.
- -representation capacity essentially reflects the model’s ability to approximate the target task’s conditional distribution, meaning that the model can approximate this distribution with a KLD no greater than .
- The assumption mainly argues the strong expressive and approximation capabilities of MLLMs. Given the strong expressive and approximation capabilities of recent large-scale MLLMs, this assumption is generally reasonable in practice.
- Moreover, numerous efforts improve the diversity of an instruction tuning dataset and robustness of the visual encoder of MLLM [1,2], which makes the learned distribution robustly approximate conditional distributions encountered during evaluation.
- As we continually pursue enriching dataset construction and enhancing visual recognition of the encoder, -representation capacity assumption becomes reasonable in more complex cases.
Second, to claim the relationship between EMID and its upper bound in a simple case (Thm 4.5), we assumed consistent conditional distributions over , , and .
- This assumption zeros out the discrepancy between conditional distributions. If the conditional distributions are quite different between ID and OOD in some real-world scenarios, this makes our upper bound underestimate the performance gap, e.g., EMID.
- However, we highlight that the strong correlations between EMID and this upper bound have been observed through 61 distribution shifts, implying the validity of our upper bound to quantify EMID.
- Meanwhile, we also provide a bound for general cases to address non-consistent conditional distributions in Thm 4.6. This general-case bound can also be empirically estimated using a procedure similar to that of Thm 4.5.
- Therefore, as mentioned in our manuscript (L302-303), we recommend choosing a proper bound based on the knowledge of the data-generating process for datasets.
A3. Details for practical implications of EMID upper bound.
While we confined the scope of this project to presenting the first theoretical framework to quantify MLLM's performance gap, we further provide a potential application of EMID upper bound, instruction tuning with regularization, for this rebuttal.
Without loss of generality, we can assume the input sequence as a sequence of intermediate representation vectors of MLLM, i.e., , and can also assume that maps this representation to responses, i.e., . This induces a modified bound with representation variable rather than raw data input .
We instantiate this modified EMID bound in two distinct setups below, where we set the 24th layer's hidden states as , and adopt RJSD [3] as a differentiable estimator for JSD and an average of empirical model output entropy. We provide evaluation results with LLaVA-v1.5-7B on in-distribution (ID) and visual (V), text (T), and joint (J) synthetic shifts.
Regularization term for instruction tuning:
- One can not access during the training phase, so we alternatively enforce the distribution of the intermediate representation to be close to the standard Gaussian.
- We sampled 10% of the instruction dataset from LLaVA-v1.5, and trained the entire LLM and modality connector parameters of LLaVA-v1.5 with and without the regularization.
- As shown in the table, we confirm that the EMID can be leveraged as a regularizer during instruction tuning to pursue better robustness to distribution shifts. |Method|ID|V Shift|T Shift|J Shift| |-|-|-|-|-| |Baseline|72.7|65.8|68.0|59.6| |Baseline + Ours|72.7|66.3|68.3|60.8|
Reference
- Qwen2.5-VL Technical Report, Alibaba Group 2025
- Eagle: Exploring The Design Space for Multimodal LLMs with Mixture of Encoders, Shi et al. 2025
- The Representation Jensen-Shannon Divergence, Hoyos-Osorio et al. 2024
I appreciate the response from the authors, which addresses my concerns. Thus, I would like to increase my score to Accept.
We thank the reviewer for taking the time to read our rebuttal and increasing the score. We appreciate your insightful comments and support!
This paper proposes Effective Mutual Information (EMI), a new metric to measure query-response relevance in MLLMs. Unlike standard Mutual Information, EMI eliminates domain-specific elements, enhancing generalizability in out-of-distribution (OOD) settings. With multiple LLaVA variants and benchmarks, experiments confirm EMI’s effectiveness, aligning with theoretical predictions across various distribution shift scenarios.